{"history_for_js":{"ydb/core/statistics/service/ut/ColumnStatistics.EqWidthHistogram":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"failure","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"failure","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"passed","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/core/statistics/service/ut/ColumnStatistics.ManyColumns":{"1773735610000000":{"status":"failure","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"failure","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723207000000":{"status":"passed","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"failure","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"failure","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"failure","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"failure","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"failure","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"failure","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"failure","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"failure","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"failure","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"failure","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"failure","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"failure","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"failure","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"failure","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"failure","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"failure","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"failure","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"failure","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"failure","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"failure","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"failure","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"failure","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"failure","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"failure","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"failure","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"failure","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"failure","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"failure","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"failure","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"failure","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"failure","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"failure","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"failure","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"failure","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"failure","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"failure","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"failure","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"failure","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"failure","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"failure","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"failure","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"failure","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"failure","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"failure","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"failure","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"failure","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"failure","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"failure","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"failure","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"failure","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"failure","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"failure","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"failure","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"failure","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"failure","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"failure","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"failure","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"failure","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"failure","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"failure","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"failure","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"failure","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773637789000000":{"status":"passed","date":"2026-03-16 05:09:49","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773637210000000":{"status":"failure","date":"2026-03-16 05:00:10","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773636520000000":{"status":"failure","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"failure","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"failure","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"failure","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"failure","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"failure","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"failure","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"failure","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"failure","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"failure","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"failure","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"failure","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"failure","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"failure","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"failure","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773550610000000":{"status":"passed","date":"2026-03-15 04:56:50","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773549981000000":{"status":"failure","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"failure","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"failure","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"failure","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"failure","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"failure","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"failure","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"failure","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"failure","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"failure","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"failure","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"failure","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"failure","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"failure","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"failure","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"failure","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"failure","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"failure","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773464285000000":{"status":"passed","date":"2026-03-14 04:58:05","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773463643000000":{"status":"failure","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"failure","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"failure","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"failure","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"failure","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"failure","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"failure","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"failure","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"failure","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"failure","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"failure","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"failure","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"failure","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"failure","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"failure","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"failure","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"failure","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"failure","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"failure","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"failure","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"failure","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"failure","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"failure","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"failure","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"failure","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"failure","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"failure","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"failure","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"failure","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"failure","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"failure","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"failure","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"failure","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"failure","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"failure","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"failure","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"failure","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"failure","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"failure","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"failure","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"failure","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"failure","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"failure","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"failure","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"failure","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"failure","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"failure","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"failure","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"failure","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"failure","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773378027000000":{"status":"passed","date":"2026-03-13 05:00:27","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773377371000000":{"status":"failure","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"failure","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"failure","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"failure","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"failure","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"failure","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"failure","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/public/sdk/cpp/src/client/topic/ut/with_direct_read_ut/Describe.DescribePartitionPermissions":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"failure","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"failure","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"passed","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"failure","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"failure","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"failure","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"failure","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"failure","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"failure","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"failure","date":"2026-03-13 16:14:44","commit":"1da63971b7460da1a9777d7d21165ea327549fa5","job_name":"PR-check","job_id":23055658532,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"failure","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"failure","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"failure","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"failure","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773386176000000":{"status":"passed","date":"2026-03-13 07:16:16","commit":"854ba0a535fa6b55e90c353f5ba524d2d16ad7f4","job_name":"Postcommit_asan","job_id":23038528519,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"},"1773360921000000":{"status":"passed","date":"2026-03-13 00:15:21","commit":"f65cf83d356f949b7d86145e279e86f49be1063b","job_name":"PR-check","job_id":23028768491,"branch":"main"}},"ydb/core/kqp/ut/sysview/KqpSystemView.Sessions":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"failure","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"passed","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773666734000000":{"status":"passed","date":"2026-03-16 13:12:14","commit":"4f0c0956c1349a03a10a76420ad9d8cc7d9ded26","job_name":"PR-check","job_id":23140508124,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773486157000000":{"status":"passed","date":"2026-03-14 11:02:37","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773426018000000":{"status":"passed","date":"2026-03-13 18:20:18","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"passed","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/core/http_proxy/ut/inside_ydb_ut/TestKinesisHttpProxy.TestListStreamConsumersWithToken":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"passed","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773666734000000":{"status":"passed","date":"2026-03-16 13:12:14","commit":"4f0c0956c1349a03a10a76420ad9d8cc7d9ded26","job_name":"PR-check","job_id":23140508124,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773656883000000":{"status":"passed","date":"2026-03-16 10:28:03","commit":"df483bc43920de8e6ed8cf2fd787c6027095e283","job_name":"PR-check","job_id":23135148620,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773486157000000":{"status":"passed","date":"2026-03-14 11:02:37","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773426018000000":{"status":"passed","date":"2026-03-13 18:20:18","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"1da63971b7460da1a9777d7d21165ea327549fa5","job_name":"PR-check","job_id":23055658532,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/public/sdk/cpp/src/client/topic/ut/with_direct_read_ut/TxUsage.Sinks_Oltp_WriteToTopic_3_Table":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"passed","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"failure","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"failure","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"failure","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773386176000000":{"status":"passed","date":"2026-03-13 07:16:16","commit":"854ba0a535fa6b55e90c353f5ba524d2d16ad7f4","job_name":"Postcommit_asan","job_id":23038528519,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"},"1773360921000000":{"status":"passed","date":"2026-03-13 00:15:21","commit":"f65cf83d356f949b7d86145e279e86f49be1063b","job_name":"PR-check","job_id":23028768491,"branch":"main"}},"yql/essentials/core/cbo/cbo_optimizer_new.cpp.clang_format":{"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773682225000000":{"status":"failure","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"}},"ydb/tests/functional/statistics/py3test.[test_analyze.py 0/10] chunk":{"1773735610000000":{"status":"failure","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"failure","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723207000000":{"status":"passed","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"failure","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"failure","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"failure","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"failure","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"failure","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"failure","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"failure","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"failure","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"failure","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"failure","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"failure","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"failure","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"failure","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"failure","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"failure","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"failure","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"failure","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"failure","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"failure","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"failure","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"failure","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"failure","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"failure","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"failure","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"failure","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"failure","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"failure","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"passed","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"failure","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"failure","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"failure","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"failure","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"failure","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"failure","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"failure","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"failure","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"failure","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"failure","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"failure","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"failure","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"failure","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"failure","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"failure","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"failure","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"failure","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"failure","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"failure","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"failure","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"failure","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"failure","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"failure","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"failure","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"failure","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"failure","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"failure","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"failure","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"failure","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"failure","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773637210000000":{"status":"passed","date":"2026-03-16 05:00:10","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773636520000000":{"status":"failure","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"failure","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"failure","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"failure","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"failure","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"failure","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"failure","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"failure","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"failure","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"failure","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"failure","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"failure","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"failure","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"failure","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"failure","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"failure","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"failure","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"failure","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"failure","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"failure","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"failure","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"failure","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"failure","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"failure","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"failure","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"failure","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"failure","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"failure","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"failure","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773464285000000":{"status":"passed","date":"2026-03-14 04:58:05","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773463643000000":{"status":"failure","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"failure","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"failure","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"failure","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"failure","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"failure","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"failure","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"failure","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"failure","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"failure","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"failure","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"failure","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"failure","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"failure","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"failure","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"failure","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"failure","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"failure","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"failure","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"failure","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"failure","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"failure","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"failure","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"failure","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"failure","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"failure","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"failure","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"failure","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"failure","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"failure","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"failure","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"failure","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"failure","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"failure","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"failure","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"failure","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"failure","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"failure","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"failure","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"failure","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"failure","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"failure","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"failure","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"failure","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"failure","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"failure","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"failure","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"failure","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"failure","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"failure","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"failure","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/datashard/truncate/concurrency/py3test.[test_truncate_table_concurrency.py 0/10] chunk":{"1773735610000000":{"status":"failure","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"failure","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"failure","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"failure","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"failure","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"failure","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"failure","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"failure","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"failure","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"failure","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"failure","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"failure","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"failure","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"failure","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"failure","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"failure","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"failure","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"failure","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"failure","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"failure","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"failure","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"failure","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"failure","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"failure","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"failure","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"failure","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"failure","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"failure","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"failure","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"failure","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"failure","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"failure","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"failure","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"failure","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"failure","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"failure","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"failure","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"failure","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"failure","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"failure","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"failure","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"failure","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"failure","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"failure","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"failure","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"failure","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"failure","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"failure","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"failure","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"failure","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"failure","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"failure","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"failure","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"failure","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"failure","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"failure","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"failure","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"failure","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"failure","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"failure","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"failure","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"failure","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"failure","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"failure","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"failure","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"failure","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"failure","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"failure","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"failure","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"failure","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"failure","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"failure","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"failure","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"failure","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"failure","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"failure","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"failure","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"failure","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"failure","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"failure","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"failure","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"failure","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"failure","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"failure","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"failure","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"failure","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"failure","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"failure","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"failure","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"failure","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"failure","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"failure","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"failure","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"failure","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/olap/scenario/test_alter_tiering.py.TestAlterTiering.test[many_tables]":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"mute","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"mute","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"mute","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"mute","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"mute","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"mute","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"mute","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"mute","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"mute","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"mute","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"mute","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"mute","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"mute","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"mute","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"mute","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773666734000000":{"status":"mute","date":"2026-03-16 13:12:14","commit":"4f0c0956c1349a03a10a76420ad9d8cc7d9ded26","job_name":"PR-check","job_id":23140508124,"branch":"main"},"1773665446000000":{"status":"mute","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"mute","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"mute","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"mute","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773658135000000":{"status":"passed","date":"2026-03-16 10:48:55","commit":"3d3cf68e68b7e2593d9482ecbd4c11ea70978da7","job_name":"PR-check","job_id":23135950903,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773645837000000":{"status":"passed","date":"2026-03-16 07:23:57","commit":"9a130331dc152f3fa65e778dbf7660ccc44414e1","job_name":"PR-check","job_id":23132123849,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773637210000000":{"status":"passed","date":"2026-03-16 05:00:10","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773636520000000":{"status":"mute","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"mute","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"mute","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"mute","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"mute","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"mute","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"mute","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488572000000":{"status":"mute","date":"2026-03-14 11:42:52","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773488352000000":{"status":"mute","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"mute","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773487377000000":{"status":"mute","date":"2026-03-14 11:22:57","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"mute","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481293000000":{"status":"passed","date":"2026-03-14 09:41:33","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Postcommit_asan","job_id":23078108567,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773479650000000":{"status":"passed","date":"2026-03-14 09:14:10","commit":"96c805ba50e2baf99fdaf1503f47767e3643287b","job_name":"PR-check","job_id":23077974662,"branch":"main"},"1773479525000000":{"status":"passed","date":"2026-03-14 09:12:05","commit":"5ef04c0596adc24aaad5217457fd21f4eaae4b21","job_name":"PR-check","job_id":23068174224,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773428448000000":{"status":"mute","date":"2026-03-13 19:00:48","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773426018000000":{"status":"mute","date":"2026-03-13 18:20:18","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"mute","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"mute","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"mute","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"mute","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"mute","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"mute","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773411419000000":{"status":"passed","date":"2026-03-13 14:16:59","commit":"6f7473e4e13f970c07a3988c2f53b61fa3aff5d2","job_name":"PR-check","job_id":23053992082,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"mute","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"passed","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"mute","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"mute","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"mute","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"mute","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"mute","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/functional/statistics/test_analyze.py.test_basic":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723207000000":{"status":"passed","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"failure","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"failure","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"failure","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"failure","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"failure","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"failure","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"failure","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"failure","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"failure","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"failure","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"failure","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"failure","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"failure","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"failure","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"failure","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"failure","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"failure","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"failure","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"failure","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"failure","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"failure","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"failure","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"failure","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"failure","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"failure","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"failure","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"failure","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"failure","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"failure","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"failure","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"failure","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"failure","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"failure","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"failure","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"failure","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"failure","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"failure","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"failure","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"failure","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"failure","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"failure","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"failure","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"failure","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"failure","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"failure","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"failure","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"failure","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"failure","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"failure","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773637210000000":{"status":"passed","date":"2026-03-16 05:00:10","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773636520000000":{"status":"failure","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"failure","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"failure","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"failure","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"failure","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"failure","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"failure","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"failure","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"failure","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"failure","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"failure","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"failure","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"failure","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"failure","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"failure","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"failure","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"failure","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"failure","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"failure","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"failure","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"failure","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"failure","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"failure","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"failure","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"failure","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"failure","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"failure","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773464285000000":{"status":"passed","date":"2026-03-14 04:58:05","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773463643000000":{"status":"failure","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"failure","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"failure","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"failure","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"failure","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"failure","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"failure","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"failure","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"failure","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"failure","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"failure","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"failure","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"failure","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"failure","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"failure","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"failure","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"failure","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"failure","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"failure","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"failure","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"failure","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"failure","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"failure","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"failure","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"failure","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"failure","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"failure","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"failure","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"failure","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"failure","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"failure","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"failure","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"failure","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"failure","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"failure","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"failure","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"failure","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"failure","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"failure","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"failure","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"failure","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"failure","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"failure","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"failure","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"failure","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"failure","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"failure","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"failure","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"failure","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"failure","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"failure","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/functional/audit/test_canonical_records.py.test_dstool_evict_vdisk_grpc":{"1773735610000000":{"status":"failure","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"failure","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723348000000":{"status":"failure","date":"2026-03-17 04:55:48","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773723207000000":{"status":"failure","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"failure","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"failure","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"failure","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"failure","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"failure","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"failure","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"failure","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"failure","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"failure","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"failure","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"failure","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"failure","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"failure","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"failure","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"failure","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"failure","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"failure","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"failure","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"failure","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"failure","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"failure","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"failure","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"failure","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"failure","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"failure","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"failure","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"failure","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"failure","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"failure","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"failure","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"failure","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"failure","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"failure","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"failure","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"failure","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"failure","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"failure","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"failure","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"failure","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"failure","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"failure","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"failure","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"failure","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"failure","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"failure","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773669292000000":{"status":"failure","date":"2026-03-16 13:54:52","commit":"4f0c0956c1349a03a10a76420ad9d8cc7d9ded26","job_name":"PR-check","job_id":23140508124,"branch":"main"},"1773668553000000":{"status":"failure","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"failure","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773668028000000":{"status":"failure","date":"2026-03-16 13:33:48","commit":"4f0c0956c1349a03a10a76420ad9d8cc7d9ded26","job_name":"PR-check","job_id":23140508124,"branch":"main"},"1773667338000000":{"status":"failure","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"failure","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"failure","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773666734000000":{"status":"failure","date":"2026-03-16 13:12:14","commit":"4f0c0956c1349a03a10a76420ad9d8cc7d9ded26","job_name":"PR-check","job_id":23140508124,"branch":"main"},"1773665446000000":{"status":"failure","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"failure","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"failure","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"failure","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"failure","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"failure","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"failure","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"failure","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"failure","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"failure","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"failure","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"failure","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773637789000000":{"status":"failure","date":"2026-03-16 05:09:49","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773637210000000":{"status":"failure","date":"2026-03-16 05:00:10","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773636520000000":{"status":"failure","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"failure","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"failure","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"failure","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"failure","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"failure","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"failure","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"failure","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"failure","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"failure","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"failure","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"failure","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"failure","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"failure","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"failure","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773550734000000":{"status":"failure","date":"2026-03-15 04:58:54","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773550610000000":{"status":"failure","date":"2026-03-15 04:56:50","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773549981000000":{"status":"failure","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"failure","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"failure","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"failure","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"failure","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"failure","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"failure","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"failure","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"failure","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"failure","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"failure","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488572000000":{"status":"failure","date":"2026-03-14 11:42:52","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773488352000000":{"status":"failure","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"failure","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773487377000000":{"status":"failure","date":"2026-03-14 11:22:57","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773486157000000":{"status":"failure","date":"2026-03-14 11:02:37","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773485812000000":{"status":"failure","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"failure","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"failure","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"failure","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"failure","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773464419000000":{"status":"failure","date":"2026-03-14 05:00:19","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773464285000000":{"status":"failure","date":"2026-03-14 04:58:05","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773463643000000":{"status":"failure","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"failure","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773428448000000":{"status":"failure","date":"2026-03-13 19:00:48","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773427248000000":{"status":"failure","date":"2026-03-13 18:40:48","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773427126000000":{"status":"failure","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773426018000000":{"status":"failure","date":"2026-03-13 18:20:18","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773423092000000":{"status":"failure","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"failure","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"failure","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"failure","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"failure","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"failure","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"failure","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"failure","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"failure","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"failure","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"failure","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"failure","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"failure","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"failure","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"failure","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"failure","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"failure","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"failure","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"failure","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"failure","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"failure","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"failure","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"failure","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"failure","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"failure","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"failure","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"failure","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"failure","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"failure","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"failure","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"failure","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"failure","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"failure","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"failure","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"failure","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"failure","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"failure","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"failure","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"failure","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"failure","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"failure","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"failure","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"failure","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"failure","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"failure","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773378178000000":{"status":"failure","date":"2026-03-13 05:02:58","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773378027000000":{"status":"failure","date":"2026-03-13 05:00:27","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773377371000000":{"status":"failure","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"failure","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"failure","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"failure","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"failure","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"failure","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"failure","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/functional/api/test_discovery.py.TestDiscoveryFaultInjectionSlotStop.test_scenario":{"1773735610000000":{"status":"mute","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723348000000":{"status":"passed","date":"2026-03-17 04:55:48","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773723207000000":{"status":"mute","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"mute","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"mute","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"mute","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"mute","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"mute","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"mute","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690345000000":{"status":"mute","date":"2026-03-16 19:45:45","commit":"8a2feac1d14fa9aaeb30300c286609ba4dfe313f","job_name":"PR-check","job_id":23161911976,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"mute","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"mute","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"mute","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"mute","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773680169000000":{"status":"passed","date":"2026-03-16 16:56:09","commit":"03434470e54820a4d64a96675d2219b33bd478a6","job_name":"PR-check","job_id":23154893676,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"mute","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"mute","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"mute","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"mute","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"mute","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"mute","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"mute","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"mute","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773666734000000":{"status":"passed","date":"2026-03-16 13:12:14","commit":"4f0c0956c1349a03a10a76420ad9d8cc7d9ded26","job_name":"PR-check","job_id":23140508124,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773661040000000":{"status":"mute","date":"2026-03-16 11:37:20","commit":"6fe8a3018f42f12d34a6eda9fdf05fedd4682934","job_name":"PR-check","job_id":23141280415,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"mute","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"mute","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"mute","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773637789000000":{"status":"mute","date":"2026-03-16 05:09:49","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773637210000000":{"status":"mute","date":"2026-03-16 05:00:10","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773636520000000":{"status":"mute","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"mute","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"mute","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"mute","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"mute","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"mute","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"mute","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"mute","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"mute","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"mute","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"mute","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773486157000000":{"status":"passed","date":"2026-03-14 11:02:37","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773485812000000":{"status":"mute","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"mute","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773426018000000":{"status":"passed","date":"2026-03-13 18:20:18","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"mute","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"1da63971b7460da1a9777d7d21165ea327549fa5","job_name":"PR-check","job_id":23055658532,"branch":"main"},"1773418466000000":{"status":"mute","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"mute","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"mute","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"mute","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"mute","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"mute","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"mute","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"mute","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"mute","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"mute","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"mute","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"mute","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"mute","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"mute","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"mute","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"mute","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"mute","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"mute","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"mute","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"mute","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"mute","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/functional/nbs/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723207000000":{"status":"passed","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"mute","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"mute","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"mute","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690345000000":{"status":"passed","date":"2026-03-16 19:45:45","commit":"8a2feac1d14fa9aaeb30300c286609ba4dfe313f","job_name":"PR-check","job_id":23161911976,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"mute","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"mute","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"mute","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773680169000000":{"status":"passed","date":"2026-03-16 16:56:09","commit":"03434470e54820a4d64a96675d2219b33bd478a6","job_name":"PR-check","job_id":23154893676,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"mute","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"mute","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"mute","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"mute","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773661040000000":{"status":"passed","date":"2026-03-16 11:37:20","commit":"6fe8a3018f42f12d34a6eda9fdf05fedd4682934","job_name":"PR-check","job_id":23141280415,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"mute","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773550734000000":{"status":"mute","date":"2026-03-15 04:58:54","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773550610000000":{"status":"mute","date":"2026-03-15 04:56:50","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773549981000000":{"status":"mute","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"mute","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"mute","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"mute","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"mute","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"mute","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"mute","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"mute","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"mute","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"mute","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773378027000000":{"status":"passed","date":"2026-03-13 05:00:27","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773377371000000":{"status":"mute","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"mute","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"mute","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"mute","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/functional/nbs/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read":{"1773735610000000":{"status":"mute","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"mute","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"mute","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"mute","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690345000000":{"status":"passed","date":"2026-03-16 19:45:45","commit":"8a2feac1d14fa9aaeb30300c286609ba4dfe313f","job_name":"PR-check","job_id":23161911976,"branch":"main"},"1773690098000000":{"status":"mute","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"mute","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"mute","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"mute","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"mute","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"mute","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773680169000000":{"status":"mute","date":"2026-03-16 16:56:09","commit":"03434470e54820a4d64a96675d2219b33bd478a6","job_name":"PR-check","job_id":23154893676,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"mute","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"mute","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"mute","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"mute","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"mute","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"mute","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"mute","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"mute","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"mute","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"mute","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"mute","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"mute","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773661040000000":{"status":"mute","date":"2026-03-16 11:37:20","commit":"6fe8a3018f42f12d34a6eda9fdf05fedd4682934","job_name":"PR-check","job_id":23141280415,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"mute","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"mute","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"mute","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"mute","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"mute","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"mute","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"mute","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"mute","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"mute","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"mute","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"mute","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"mute","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"mute","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"mute","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"mute","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"mute","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"mute","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"mute","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"mute","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773464419000000":{"status":"passed","date":"2026-03-14 05:00:19","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773464285000000":{"status":"mute","date":"2026-03-14 04:58:05","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773463643000000":{"status":"mute","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"mute","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"mute","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"mute","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"mute","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"mute","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"mute","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"mute","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"mute","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"mute","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"mute","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"mute","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"mute","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"mute","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"mute","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"mute","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"mute","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"mute","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"}},"ydb/tests/functional/statistics/test_restarts.py.test_basic":{"1773735610000000":{"status":"mute","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"mute","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723348000000":{"status":"passed","date":"2026-03-17 04:55:48","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773723207000000":{"status":"mute","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"mute","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"mute","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"mute","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"mute","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"mute","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"mute","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"mute","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"mute","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"mute","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"mute","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"mute","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"mute","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"mute","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"mute","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"mute","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"mute","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"mute","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"mute","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"mute","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"mute","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"mute","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"mute","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"mute","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"mute","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"mute","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"mute","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"mute","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"mute","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"mute","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"mute","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"mute","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"mute","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"mute","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"mute","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"mute","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"mute","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"mute","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"mute","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"mute","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"mute","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"mute","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"mute","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"mute","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"mute","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"mute","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"mute","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"mute","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"mute","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"mute","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"mute","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"mute","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"mute","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"mute","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"mute","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"mute","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"mute","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"mute","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"mute","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"mute","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"mute","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"mute","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"mute","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"mute","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"mute","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"mute","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"mute","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"mute","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"mute","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773637210000000":{"status":"passed","date":"2026-03-16 05:00:10","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773636520000000":{"status":"mute","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"mute","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"mute","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"mute","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"mute","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"mute","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"mute","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"mute","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"mute","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"mute","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"mute","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"mute","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"mute","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"mute","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"mute","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773550610000000":{"status":"passed","date":"2026-03-15 04:56:50","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773549981000000":{"status":"mute","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"mute","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"mute","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"mute","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"mute","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"mute","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"mute","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"mute","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"mute","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"mute","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"mute","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"mute","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"mute","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"mute","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"mute","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"mute","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"mute","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"mute","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773464285000000":{"status":"passed","date":"2026-03-14 04:58:05","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773463643000000":{"status":"mute","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"mute","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"mute","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"mute","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"mute","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"mute","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"mute","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"mute","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"mute","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"mute","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"mute","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"mute","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"mute","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"mute","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"mute","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"mute","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"mute","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"mute","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"mute","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"mute","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"mute","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"mute","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"mute","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"mute","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"mute","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"mute","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"mute","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"mute","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"mute","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"mute","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"mute","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"mute","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"mute","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"mute","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"mute","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"mute","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"mute","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"mute","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"mute","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"mute","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"mute","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"mute","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"mute","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"mute","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"mute","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"mute","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"mute","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"mute","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"mute","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"mute","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773378178000000":{"status":"passed","date":"2026-03-13 05:02:58","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773378027000000":{"status":"mute","date":"2026-03-13 05:00:27","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773377371000000":{"status":"mute","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"mute","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"mute","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"mute","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"mute","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"mute","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"mute","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/tools/pq_read/test/test_timeout.py.TestTimeout.test_timeout":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"failure","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"failure","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"failure","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"failure","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"failure","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773666734000000":{"status":"passed","date":"2026-03-16 13:12:14","commit":"4f0c0956c1349a03a10a76420ad9d8cc7d9ded26","job_name":"PR-check","job_id":23140508124,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"failure","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"failure","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"failure","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"failure","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"failure","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"failure","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"failure","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773486157000000":{"status":"passed","date":"2026-03-14 11:02:37","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"failure","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"failure","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"failure","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"failure","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773426018000000":{"status":"passed","date":"2026-03-13 18:20:18","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"failure","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"failure","date":"2026-03-13 16:14:44","commit":"1da63971b7460da1a9777d7d21165ea327549fa5","job_name":"PR-check","job_id":23055658532,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"failure","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"failure","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/datashard/ttl/test_ttl.py.TestTTL.test_ttl[table_Date_1_UNIQUE_SYNC-pk_types35-all_types35-index35-Date-UNIQUE-SYNC]":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723207000000":{"status":"passed","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"failure","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"failure","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"failure","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"failure","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"failure","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"failure","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"failure","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"failure","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"failure","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"failure","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"failure","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"failure","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"failure","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"failure","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"failure","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"failure","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"failure","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"failure","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"failure","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"failure","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"failure","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"failure","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"failure","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"failure","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"failure","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"failure","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"failure","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"failure","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"failure","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"failure","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"failure","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"failure","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"failure","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"failure","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"passed","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"failure","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"failure","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"failure","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"failure","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"failure","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"failure","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"failure","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"failure","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"failure","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"failure","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"failure","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"failure","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"failure","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"failure","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"failure","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"failure","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"failure","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"failure","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"failure","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"failure","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"failure","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/tests/stress/kv/tests/test_workload.py.TestYdbKvWorkload.test[column]":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773706663000000":{"status":"passed","date":"2026-03-17 00:17:43","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-whitelist-run","job_id":23170162206,"branch":"main"},"1773704267000000":{"status":"passed","date":"2026-03-16 23:37:47","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_stress","job_id":23170211798,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773666734000000":{"status":"passed","date":"2026-03-16 13:12:14","commit":"4f0c0956c1349a03a10a76420ad9d8cc7d9ded26","job_name":"PR-check","job_id":23140508124,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773620795000000":{"status":"passed","date":"2026-03-16 00:26:35","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-whitelist-run","job_id":23121458291,"branch":"main"},"1773619671000000":{"status":"passed","date":"2026-03-16 00:07:51","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_stress","job_id":23121485170,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773533668000000":{"status":"passed","date":"2026-03-15 00:14:28","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-whitelist-run","job_id":23098331199,"branch":"main"},"1773531294000000":{"status":"passed","date":"2026-03-14 23:34:54","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_stress","job_id":23098357422,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"failure","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773486157000000":{"status":"passed","date":"2026-03-14 11:02:37","commit":"f039416ddfe3be8445b874b2801e5895f7f1fa4c","job_name":"PR-check","job_id":23064393528,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773447423000000":{"status":"passed","date":"2026-03-14 00:17:03","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Regression-whitelist-run","job_id":23074055774,"branch":"main"},"1773445519000000":{"status":"passed","date":"2026-03-13 23:45:19","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Regression-run_stress","job_id":23074097319,"branch":"main"},"1773445080000000":{"status":"failure","date":"2026-03-13 23:38:00","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Regression-run_stress","job_id":23074097319,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773426018000000":{"status":"passed","date":"2026-03-13 18:20:18","commit":"7504da4ce8935ef8a45844f2b4edb37b6edf4a86","job_name":"PR-check","job_id":23060060351,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"1da63971b7460da1a9777d7d21165ea327549fa5","job_name":"PR-check","job_id":23055658532,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"},"1773361143000000":{"status":"passed","date":"2026-03-13 00:19:03","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-whitelist-run","job_id":23028098603,"branch":"main"}},"ydb/tests/stress/topic/tests/test_workload_topic.py.TestYdbTopicWorkload.test":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"failure","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723207000000":{"status":"passed","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"failure","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773706663000000":{"status":"passed","date":"2026-03-17 00:17:43","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-whitelist-run","job_id":23170162206,"branch":"main"},"1773704267000000":{"status":"passed","date":"2026-03-16 23:37:47","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_stress","job_id":23170211798,"branch":"main"},"1773702149000000":{"status":"failure","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"failure","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"failure","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"failure","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"failure","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"failure","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"failure","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"failure","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"failure","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"failure","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"failure","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"failure","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"failure","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"failure","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"failure","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"failure","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"failure","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"failure","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"failure","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"failure","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"failure","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"failure","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"failure","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"failure","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"failure","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"failure","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"failure","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"failure","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"failure","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"failure","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"failure","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"failure","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"failure","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"failure","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"failure","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"failure","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"failure","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"failure","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"failure","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"failure","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"failure","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"failure","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"failure","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"failure","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"failure","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"failure","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"failure","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"failure","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"failure","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"failure","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"failure","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"failure","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"failure","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"failure","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"failure","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"failure","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773647899000000":{"status":"failure","date":"2026-03-16 07:58:19","commit":"580e7354c64d6e5cf10c41b7f4323767366ab144","job_name":"PR-check","job_id":23133232992,"branch":"main"},"1773644268000000":{"status":"failure","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"failure","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773637789000000":{"status":"passed","date":"2026-03-16 05:09:49","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773637210000000":{"status":"failure","date":"2026-03-16 05:00:10","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773636520000000":{"status":"failure","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"failure","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773628100000000":{"status":"failure","date":"2026-03-16 02:28:20","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-whitelist-run","job_id":23121458291,"branch":"main"},"1773627098000000":{"status":"failure","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"failure","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773624453000000":{"status":"failure","date":"2026-03-16 01:27:33","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-whitelist-run","job_id":23121458291,"branch":"main"},"1773622583000000":{"status":"failure","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773620795000000":{"status":"failure","date":"2026-03-16 00:26:35","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-whitelist-run","job_id":23121458291,"branch":"main"},"1773619671000000":{"status":"passed","date":"2026-03-16 00:07:51","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_stress","job_id":23121485170,"branch":"main"},"1773617923000000":{"status":"failure","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"failure","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"failure","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"failure","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"failure","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"failure","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"failure","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"failure","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"failure","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"failure","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773550734000000":{"status":"passed","date":"2026-03-15 04:58:54","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773550610000000":{"status":"failure","date":"2026-03-15 04:56:50","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773549981000000":{"status":"failure","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"failure","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773537307000000":{"status":"passed","date":"2026-03-15 01:15:07","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-whitelist-run","job_id":23098331199,"branch":"main"},"1773533668000000":{"status":"failure","date":"2026-03-15 00:14:28","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-whitelist-run","job_id":23098331199,"branch":"main"},"1773532000000000":{"status":"failure","date":"2026-03-14 23:46:40","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_stress","job_id":23098357422,"branch":"main"},"1773531735000000":{"status":"failure","date":"2026-03-14 23:42:15","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_stress","job_id":23098357422,"branch":"main"},"1773531294000000":{"status":"failure","date":"2026-03-14 23:34:54","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_stress","job_id":23098357422,"branch":"main"},"1773526956000000":{"status":"failure","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"failure","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"failure","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"failure","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"failure","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"failure","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"failure","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"failure","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"failure","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"failure","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"failure","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"failure","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"failure","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"failure","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"failure","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"failure","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773464419000000":{"status":"failure","date":"2026-03-14 05:00:19","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773464285000000":{"status":"failure","date":"2026-03-14 04:58:05","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773463643000000":{"status":"failure","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773454713000000":{"status":"passed","date":"2026-03-14 02:18:33","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Regression-whitelist-run","job_id":23074055774,"branch":"main"},"1773451079000000":{"status":"failure","date":"2026-03-14 01:17:59","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Regression-whitelist-run","job_id":23074055774,"branch":"main"},"1773447423000000":{"status":"failure","date":"2026-03-14 00:17:03","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Regression-whitelist-run","job_id":23074055774,"branch":"main"},"1773445080000000":{"status":"passed","date":"2026-03-13 23:38:00","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Regression-run_stress","job_id":23074097319,"branch":"main"},"1773428652000000":{"status":"failure","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"failure","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"failure","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"failure","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"failure","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"failure","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"failure","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"failure","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"failure","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"failure","date":"2026-03-13 16:14:44","commit":"1da63971b7460da1a9777d7d21165ea327549fa5","job_name":"PR-check","job_id":23055658532,"branch":"main"},"1773418466000000":{"status":"failure","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"failure","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"failure","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"failure","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"failure","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"failure","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"failure","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"failure","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"failure","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"failure","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"failure","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"failure","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"failure","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"failure","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"failure","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"failure","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"failure","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"failure","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"failure","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"failure","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"failure","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"failure","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"failure","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"failure","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"failure","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"failure","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"failure","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"failure","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"failure","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"failure","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"failure","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"failure","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"failure","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389368000000":{"status":"failure","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"failure","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"failure","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773389088000000":{"status":"failure","date":"2026-03-13 08:04:48","commit":"b0604ef768605680262d453a94dbd52732fe7dc1","job_name":"PR-check","job_id":23041631592,"branch":"main"},"1773388600000000":{"status":"failure","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773378178000000":{"status":"passed","date":"2026-03-13 05:02:58","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773378027000000":{"status":"failure","date":"2026-03-13 05:00:27","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773377371000000":{"status":"failure","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"failure","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"failure","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"failure","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"failure","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364790000000":{"status":"passed","date":"2026-03-13 01:19:50","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-whitelist-run","job_id":23028098603,"branch":"main"},"1773364460000000":{"status":"failure","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"failure","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"},"1773361143000000":{"status":"failure","date":"2026-03-13 00:19:03","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-whitelist-run","job_id":23028098603,"branch":"main"}},"ydb/core/statistics/service/ut/unittest.[21/60] chunk":{"1773735610000000":{"status":"failure","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"failure","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773723207000000":{"status":"passed","date":"2026-03-17 04:53:27","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773722559000000":{"status":"failure","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"failure","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"failure","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"failure","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"failure","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"failure","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"failure","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"failure","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"failure","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"failure","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"failure","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"failure","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"failure","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"failure","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"failure","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"failure","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"failure","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"failure","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"failure","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"failure","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"failure","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"failure","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"failure","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"failure","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"failure","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"failure","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"failure","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"failure","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"failure","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"failure","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"passed","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"failure","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"failure","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"failure","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"failure","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"failure","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"failure","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"failure","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"failure","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"failure","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"failure","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"failure","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"failure","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"failure","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"failure","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"failure","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"failure","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"failure","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"failure","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"failure","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"failure","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"failure","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"failure","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"failure","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"failure","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"failure","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"failure","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"failure","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"failure","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"failure","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"failure","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"failure","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"failure","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773637789000000":{"status":"passed","date":"2026-03-16 05:09:49","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773637210000000":{"status":"passed","date":"2026-03-16 05:00:10","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773636520000000":{"status":"failure","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"failure","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"failure","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"failure","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"failure","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"failure","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"failure","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"failure","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"failure","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"failure","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"failure","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"failure","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"failure","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"failure","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"failure","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773550610000000":{"status":"passed","date":"2026-03-15 04:56:50","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773549981000000":{"status":"failure","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"failure","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"failure","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"failure","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"failure","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"failure","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"failure","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"failure","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"failure","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"failure","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"failure","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"failure","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"failure","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"failure","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"failure","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"failure","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"failure","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"failure","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773464285000000":{"status":"passed","date":"2026-03-14 04:58:05","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773463643000000":{"status":"failure","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"failure","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"failure","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"failure","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"failure","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"failure","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"failure","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"failure","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"failure","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"failure","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"failure","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"failure","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"failure","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"failure","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"failure","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"failure","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"failure","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"failure","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"failure","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"failure","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"failure","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"failure","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"failure","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"failure","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"failure","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"failure","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"failure","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"failure","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"failure","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"failure","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"failure","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"failure","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773405071000000":{"status":"failure","date":"2026-03-13 12:31:11","commit":"7bf789f021c65b85253b43e3f3179bbe5e5eab0d","job_name":"Postcommit_asan","job_id":23048034481,"branch":"main"},"1773404601000000":{"status":"failure","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"failure","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"failure","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"failure","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"failure","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"failure","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"failure","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"failure","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"failure","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"failure","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"failure","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"failure","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"failure","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"failure","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"failure","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"failure","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773378027000000":{"status":"passed","date":"2026-03-13 05:00:27","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773377371000000":{"status":"failure","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"failure","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"failure","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"failure","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"failure","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"failure","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"failure","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/core/tx/sequenceproxy/ut/unittest.sole chunk":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"failure","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"passed","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/core/graph/shard/ut/unittest.sole chunk":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"failure","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"failure","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"failure","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"passed","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"failure","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"failure","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/core/tx/sequenceshard/ut/unittest.sole chunk":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"passed","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"passed","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"passed","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"passed","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/core/tx/schemeshard/ut_truncate_table_simple/unittest.sole chunk":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"failure","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"passed","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"passed","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"failure","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"passed","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"passed","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"passed","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"passed","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"passed","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"failure","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"failure","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"passed","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"passed","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"passed","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"failure","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"passed","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"failure","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"failure","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"passed","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"failure","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773658135000000":{"status":"passed","date":"2026-03-16 10:48:55","commit":"3d3cf68e68b7e2593d9482ecbd4c11ea70978da7","job_name":"PR-check","job_id":23135950903,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"failure","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"failure","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"failure","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"passed","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"passed","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"failure","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"failure","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"passed","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"passed","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}},"ydb/core/ymq/actor/cloud_events/cloud_events_ut/unittest.sole chunk":{"1773735610000000":{"status":"passed","date":"2026-03-17 08:20:10","commit":"3792df182f021078f499e3aff4dca08c05f66457","job_name":"PR-check","job_id":23181876562,"branch":"main"},"1773729526000000":{"status":"passed","date":"2026-03-17 06:38:46","commit":"f294f2cd25a9b5833995ee6155dd1a49576f33ac","job_name":"PR-check","job_id":23179099117,"branch":"main"},"1773722559000000":{"status":"passed","date":"2026-03-17 04:42:39","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Regression-run_Small_and_Medium","job_id":23171227078,"branch":"main"},"1773720290000000":{"status":"passed","date":"2026-03-17 04:04:50","commit":"4204b0cdd9e93f9522a6909f3a63873d0326aa33","job_name":"PR-check","job_id":23173899290,"branch":"main"},"1773719674000000":{"status":"passed","date":"2026-03-17 03:54:34","commit":"8881b558e18f09db4369c7de22cee8f3184af9aa","job_name":"PR-check","job_id":23173381735,"branch":"main"},"1773717082000000":{"status":"failure","date":"2026-03-17 03:11:22","commit":"010fa78dbe12a11bb9d2ccbfa6011ec94cfe74ab","job_name":"PR-check","job_id":23173298174,"branch":"main"},"1773708708000000":{"status":"failure","date":"2026-03-17 00:51:48","commit":"4006a62dc3d244e0419327d9cbd5979310d1cb32","job_name":"PR-check","job_id":23169988380,"branch":"main"},"1773702149000000":{"status":"passed","date":"2026-03-16 23:02:29","commit":"aeff9430b8901290ee83a5ca1322d1233a016877","job_name":"PR-check","job_id":23165157486,"branch":"main"},"1773702129000000":{"status":"failure","date":"2026-03-16 23:02:09","commit":"4763eda82c95242215148bff191262887baa1e26","job_name":"PR-check","job_id":23164499557,"branch":"main"},"1773698398000000":{"status":"passed","date":"2026-03-16 21:59:58","commit":"59c5b07fc52bfc4e047331be4b65156c9303227a","job_name":"Postcommit_asan","job_id":23163009237,"branch":"main"},"1773694511000000":{"status":"passed","date":"2026-03-16 20:55:11","commit":"e830c66fced3da2f75fb97bfbf0280212c53c62c","job_name":"Postcommit_asan","job_id":23161497388,"branch":"main"},"1773692039000000":{"status":"passed","date":"2026-03-16 20:13:59","commit":"159c3c6fe513b8cc403aed3240ec34090dd31a06","job_name":"PR-check","job_id":23157647621,"branch":"main"},"1773690853000000":{"status":"failure","date":"2026-03-16 19:54:13","commit":"e27b7d91eae32acccf8a6e113c1de47538599ef6","job_name":"Postcommit_asan","job_id":23158772036,"branch":"main"},"1773690441000000":{"status":"passed","date":"2026-03-16 19:47:21","commit":"1a3ba8273bc41252fab7fa43c413081742cf6dbe","job_name":"PR-check","job_id":23156606125,"branch":"main"},"1773690098000000":{"status":"passed","date":"2026-03-16 19:41:38","commit":"a491e27fdd122067c1df09ec2d617b42c298a874","job_name":"PR-check","job_id":23156043623,"branch":"main"},"1773689559000000":{"status":"passed","date":"2026-03-16 19:32:39","commit":"29043718a3b2f1de84c09796778c447d9e2158a4","job_name":"PR-check","job_id":23158227739,"branch":"main"},"1773689025000000":{"status":"passed","date":"2026-03-16 19:23:45","commit":"61f5a6a4c9ab13ca89dd6d7988213558186c0aac","job_name":"PR-check","job_id":23157731332,"branch":"main"},"1773688476000000":{"status":"failure","date":"2026-03-16 19:14:36","commit":"92c4030dff8ca72612f332a0e8c719cdd4c955a2","job_name":"Postcommit_asan","job_id":23157009604,"branch":"main"},"1773687448000000":{"status":"failure","date":"2026-03-16 18:57:28","commit":"69dd3e3daa9c4578f5478939cb484cb03c4a6e86","job_name":"Postcommit_asan","job_id":23155360481,"branch":"main"},"1773687172000000":{"status":"passed","date":"2026-03-16 18:52:52","commit":"18576d8bb3f03ea7624d4ef77494eb9d16bc0ee3","job_name":"PR-check","job_id":23155511022,"branch":"main"},"1773687167000000":{"status":"failure","date":"2026-03-16 18:52:47","commit":"266af0ef30a5305ddaed99197749bd2a57fb0a2e","job_name":"Postcommit_asan","job_id":23155174036,"branch":"main"},"1773687158000000":{"status":"passed","date":"2026-03-16 18:52:38","commit":"13e807cc6215267c15e73389829023fb4d2afd0b","job_name":"PR-check","job_id":23155478569,"branch":"main"},"1773687146000000":{"status":"passed","date":"2026-03-16 18:52:26","commit":"b20af686c45b64437d1192098b59c8232f1cdc79","job_name":"Postcommit_asan","job_id":23155306877,"branch":"main"},"1773685622000000":{"status":"passed","date":"2026-03-16 18:27:02","commit":"19a585d71489a1e08c5e197a52c7da50ce25a64f","job_name":"PR-check","job_id":23154071291,"branch":"main"},"1773685521000000":{"status":"passed","date":"2026-03-16 18:25:21","commit":"327a215f52f78edee9477f77191344566a092b33","job_name":"Postcommit_asan","job_id":23153887084,"branch":"main"},"1773683098000000":{"status":"passed","date":"2026-03-16 17:44:58","commit":"25833b3198fae2e19cf2602351c8806137275402","job_name":"PR-check","job_id":23152810045,"branch":"main"},"1773682225000000":{"status":"passed","date":"2026-03-16 17:30:25","commit":"5ce5aeddae641a91d9cf36d3f4da4af663599333","job_name":"PR-check","job_id":23151903696,"branch":"main"},"1773682040000000":{"status":"passed","date":"2026-03-16 17:27:20","commit":"ff581f423a10e4cdda12ebbcffc799d4322f18da","job_name":"PR-check","job_id":23150243033,"branch":"main"},"1773681818000000":{"status":"failure","date":"2026-03-16 17:23:38","commit":"5dc1fb0393cc51156f6f43f709a81dd1f7aad471","job_name":"PR-check","job_id":23151499975,"branch":"main"},"1773680250000000":{"status":"passed","date":"2026-03-16 16:57:30","commit":"a40a620a397b7848560a75b2ab8abe2fa3e47ec2","job_name":"Postcommit_asan","job_id":23150046827,"branch":"main"},"1773679933000000":{"status":"passed","date":"2026-03-16 16:52:13","commit":"f8eecf922c42574f5195eab271594b113aba3de7","job_name":"PR-check","job_id":23150715061,"branch":"main"},"1773679449000000":{"status":"passed","date":"2026-03-16 16:44:09","commit":"f30f0c4ca7e4432919a088bd849ce5d4910614d1","job_name":"PR-check","job_id":23150470399,"branch":"main"},"1773679120000000":{"status":"passed","date":"2026-03-16 16:38:40","commit":"e1fe12222f2b4647e4e6263b93ca61e6a4319e01","job_name":"PR-check","job_id":23148070884,"branch":"main"},"1773679007000000":{"status":"passed","date":"2026-03-16 16:36:47","commit":"e5b992f6a6391dbe8617a8e18f77c121b883ed44","job_name":"PR-check","job_id":23147492794,"branch":"main"},"1773678560000000":{"status":"passed","date":"2026-03-16 16:29:20","commit":"628d0ae5f3db23abb186cb3e1778d36752cebc5e","job_name":"PR-check","job_id":23149495539,"branch":"main"},"1773678541000000":{"status":"failure","date":"2026-03-16 16:29:01","commit":"39371c54c2322f3e6144d9d7e4cbd097578a8533","job_name":"PR-check","job_id":23149387953,"branch":"main"},"1773678224000000":{"status":"passed","date":"2026-03-16 16:23:44","commit":"17717eaa5b5ddb78699117e34aadbd43a23af657","job_name":"PR-check","job_id":23149375630,"branch":"main"},"1773677952000000":{"status":"failure","date":"2026-03-16 16:19:12","commit":"a3ddb1b7a7b1f9b1489549d2615987f20a6a4943","job_name":"PR-check","job_id":23149197580,"branch":"main"},"1773677718000000":{"status":"passed","date":"2026-03-16 16:15:18","commit":"bf0b488a0814e63568eb754683d79f2c46ed4278","job_name":"PR-check","job_id":23148669165,"branch":"main"},"1773677510000000":{"status":"passed","date":"2026-03-16 16:11:50","commit":"0d420491cfaec9e442cc4da153a3d42cd7d00a53","job_name":"PR-check","job_id":23149029900,"branch":"main"},"1773677230000000":{"status":"failure","date":"2026-03-16 16:07:10","commit":"ba11d3d79642ab639e3de2547df935345dd06f86","job_name":"PR-check","job_id":23148072664,"branch":"main"},"1773677073000000":{"status":"passed","date":"2026-03-16 16:04:33","commit":"5e505b5bf44e9655b84f4690ad7431e6422d959b","job_name":"Postcommit_asan","job_id":23148384404,"branch":"main"},"1773676288000000":{"status":"passed","date":"2026-03-16 15:51:28","commit":"35c364d992efdbdda06df93009166eccb0f29c20","job_name":"Postcommit_asan","job_id":23147058527,"branch":"main"},"1773675427000000":{"status":"failure","date":"2026-03-16 15:37:07","commit":"0a41772fbb6f40fc0ba2572975bcf494d0d5b8c1","job_name":"PR-check","job_id":23145222125,"branch":"main"},"1773675081000000":{"status":"passed","date":"2026-03-16 15:31:21","commit":"abbca7feb91d7b15908033502e14a929a6e4e203","job_name":"Postcommit_asan","job_id":23146012754,"branch":"main"},"1773674966000000":{"status":"passed","date":"2026-03-16 15:29:26","commit":"4ce82ba977f317460dcedc45006648b5ddca79a5","job_name":"PR-check","job_id":23146299041,"branch":"main"},"1773674896000000":{"status":"passed","date":"2026-03-16 15:28:16","commit":"0bf016cd3ba3a0aaec0490526b0b612a20fa2f30","job_name":"PR-check","job_id":23145753558,"branch":"main"},"1773674709000000":{"status":"passed","date":"2026-03-16 15:25:09","commit":"edb055a1ce29d71bcf436f77294431c4dfa887dd","job_name":"Postcommit_asan","job_id":23146510402,"branch":"main"},"1773674432000000":{"status":"passed","date":"2026-03-16 15:20:32","commit":"e52cceb8aea4b56eafb6f71dc278fe545ceb6248","job_name":"Postcommit_asan","job_id":23145869520,"branch":"main"},"1773672978000000":{"status":"passed","date":"2026-03-16 14:56:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23148582725,"branch":"main"},"1773672747000000":{"status":"passed","date":"2026-03-16 14:52:27","commit":"fb89e3d2894067d77685bf211a0a9d78d795e0bf","job_name":"PR-check","job_id":23145096195,"branch":"main"},"1773671187000000":{"status":"failure","date":"2026-03-16 14:26:27","commit":"b6066a55caacfa73a29fa2b44b7dc59ba05d1580","job_name":"PR-check","job_id":23144064128,"branch":"main"},"1773668553000000":{"status":"passed","date":"2026-03-16 13:42:33","commit":"a51bae76a1ea2eca2aab5351d5b1c86f3850bec6","job_name":"Postcommit_asan","job_id":23145419648,"branch":"main"},"1773668400000000":{"status":"passed","date":"2026-03-16 13:40:00","commit":"b8c37888fc86751000e22879dd2a1ce337a1d743","job_name":"PR-check","job_id":23142544167,"branch":"main"},"1773667338000000":{"status":"failure","date":"2026-03-16 13:22:18","commit":"74a4d64177006345a5d3fa7f690a7a0f17f95fb0","job_name":"PR-check","job_id":23141173396,"branch":"main"},"1773667038000000":{"status":"passed","date":"2026-03-16 13:17:18","commit":"88bed7837e5b74cc3d960052ad668d06a47f5294","job_name":"PR-check","job_id":23141683154,"branch":"main"},"1773666791000000":{"status":"passed","date":"2026-03-16 13:13:11","commit":"e3c0e419584cd6687e5d9c29e68037ad218f026a","job_name":"PR-check","job_id":23141698957,"branch":"main"},"1773665446000000":{"status":"passed","date":"2026-03-16 12:50:46","commit":"60b67decf312eaf05b06a154e0ee31caf41db69b","job_name":"PR-check","job_id":23140565707,"branch":"main"},"1773663833000000":{"status":"passed","date":"2026-03-16 12:23:53","commit":"aac10da265d11054ae97f55db994b6084cfc219e","job_name":"PR-check","job_id":23139520641,"branch":"main"},"1773661068000000":{"status":"passed","date":"2026-03-16 11:37:48","commit":"412b4dd71ab8394eda1b2add760360d7b938349f","job_name":"PR-check","job_id":23137650572,"branch":"main"},"1773660879000000":{"status":"failure","date":"2026-03-16 11:34:39","commit":"e83cae2f1d3ca3f899916833d64b589e0b5614bc","job_name":"PR-check","job_id":23137361653,"branch":"main"},"1773659987000000":{"status":"passed","date":"2026-03-16 11:19:47","commit":"bc3c6a362339c94e95b598dca56b1efeeb0a4fe8","job_name":"PR-check","job_id":23137020911,"branch":"main"},"1773659879000000":{"status":"passed","date":"2026-03-16 11:17:59","commit":"bdd693c8bd23d7f276744abc68d4efa10cda3bd3","job_name":"PR-check","job_id":23137065097,"branch":"main"},"1773655409000000":{"status":"passed","date":"2026-03-16 10:03:29","commit":"b82952457f687403af1836e8f481e625bb40cfb0","job_name":"PR-check","job_id":23132834548,"branch":"main"},"1773653768000000":{"status":"passed","date":"2026-03-16 09:36:08","commit":"5e2697d38803994c283a20430c44e6057fb4909d","job_name":"Postcommit_asan","job_id":23132255188,"branch":"main"},"1773653670000000":{"status":"passed","date":"2026-03-16 09:34:30","commit":"c1ea51abd05fd9db08abd731371d22cd9087d239","job_name":"PR-check","job_id":23133271993,"branch":"main"},"1773651160000000":{"status":"passed","date":"2026-03-16 08:52:40","commit":"3b58deeb539eeabd9466745038c8a633704f5b68","job_name":"PR-check","job_id":23131917867,"branch":"main"},"1773644268000000":{"status":"passed","date":"2026-03-16 06:57:48","commit":"ed85fc2bde222905c022990c3615230669cd6ad8","job_name":"Postcommit_asan","job_id":23128309762,"branch":"main"},"1773643731000000":{"status":"passed","date":"2026-03-16 06:48:51","commit":"80c750288a330e2c3e34c5dc48f29fb9072c6904","job_name":"Postcommit_asan","job_id":23128432571,"branch":"main"},"1773642105000000":{"status":"passed","date":"2026-03-16 06:21:45","commit":"d42b2839e80061f69756cc4e4c067e9b7f44e30a","job_name":"Postcommit_asan","job_id":23128223000,"branch":"main"},"1773636520000000":{"status":"passed","date":"2026-03-16 04:48:40","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Regression-run_Small_and_Medium","job_id":23122080260,"branch":"main"},"1773634122000000":{"status":"passed","date":"2026-03-16 04:08:42","commit":"7d8f61db8d3e93a45ba674de5912bb3f170c65a1","job_name":"PR-check","job_id":23123634838,"branch":"main"},"1773627098000000":{"status":"passed","date":"2026-03-16 02:11:38","commit":"96a73dd0057ca5f1c80787a8590dc41d7cf0c952","job_name":"PR-check","job_id":23121989332,"branch":"main"},"1773626829000000":{"status":"passed","date":"2026-03-16 02:07:09","commit":"280a117bc135cce8d33ea9e85e39cb905cacefaa","job_name":"PR-check","job_id":23122098688,"branch":"main"},"1773622583000000":{"status":"passed","date":"2026-03-16 00:56:23","commit":"814075dd2766ea47577413e15464fdfdb6eda6f9","job_name":"PR-check","job_id":23121576332,"branch":"main"},"1773617923000000":{"status":"passed","date":"2026-03-15 23:38:43","commit":"908699294519aba78a6fbe59c58ae0b3d8cbdc8f","job_name":"PR-check","job_id":23120216774,"branch":"main"},"1773615765000000":{"status":"passed","date":"2026-03-15 23:02:45","commit":"dc0ec714f17c1a8dba32fdd30df78903b2986529","job_name":"PR-check","job_id":23118700341,"branch":"main"},"1773615308000000":{"status":"passed","date":"2026-03-15 22:55:08","commit":"376b0ce7362a7a4ad16667c51950a52e9d0d4cf8","job_name":"PR-check","job_id":23119310582,"branch":"main"},"1773608284000000":{"status":"passed","date":"2026-03-15 20:58:04","commit":"83c1d8a6c87b177e79dbcdb9061f10a17d6b58bb","job_name":"PR-check","job_id":23117284380,"branch":"main"},"1773606561000000":{"status":"passed","date":"2026-03-15 20:29:21","commit":"4b4add5f8611db9355d3228cc2ece8cbb34ce865","job_name":"PR-check","job_id":23115983920,"branch":"main"},"1773592886000000":{"status":"passed","date":"2026-03-15 16:41:26","commit":"e0240123fed3f537395028e548f56f21b1aed34e","job_name":"Postcommit_asan","job_id":23113959180,"branch":"main"},"1773572070000000":{"status":"passed","date":"2026-03-15 10:54:30","commit":"54574ce552bdc86fc57f0f65e6ddeaeacfb08511","job_name":"PR-check","job_id":23107363803,"branch":"main"},"1773570977000000":{"status":"passed","date":"2026-03-15 10:36:17","commit":"6cba82646a194055dd515307baf33eb6963078b0","job_name":"PR-check","job_id":23107021843,"branch":"main"},"1773562367000000":{"status":"passed","date":"2026-03-15 08:12:47","commit":"5136d26e8e630a42da748b8adb65e39cb3d3da81","job_name":"PR-check","job_id":23104894790,"branch":"main"},"1773562210000000":{"status":"passed","date":"2026-03-15 08:10:10","commit":"a83a5130d84e4bfdbf387089440a5a1b9a69ab80","job_name":"Postcommit_asan","job_id":23105945011,"branch":"main"},"1773549981000000":{"status":"passed","date":"2026-03-15 04:46:21","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Regression-run_Small_and_Medium","job_id":23098915538,"branch":"main"},"1773547892000000":{"status":"passed","date":"2026-03-15 04:11:32","commit":"f948cc7c702bf3ef9ea3e0a2309dab0a8e2f90ef","job_name":"PR-check","job_id":23100244917,"branch":"main"},"1773526956000000":{"status":"passed","date":"2026-03-14 22:22:36","commit":"fadceb32d63fb122f219906ad4afebbd06e9fdf8","job_name":"PR-check","job_id":23095216129,"branch":"main"},"1773519197000000":{"status":"passed","date":"2026-03-14 20:13:17","commit":"f22b5af5e39be920e0da12e1bd8c95fbf136b4b6","job_name":"PR-check","job_id":23093608985,"branch":"main"},"1773517997000000":{"status":"passed","date":"2026-03-14 19:53:17","commit":"39c458f2d8a010abbd80ba58f409f1c74e4e40c3","job_name":"PR-check","job_id":23092877215,"branch":"main"},"1773510184000000":{"status":"passed","date":"2026-03-14 17:43:04","commit":"407e7fec419e06989afa5f81618a90dd586dd4bc","job_name":"PR-check","job_id":23091276471,"branch":"main"},"1773508856000000":{"status":"passed","date":"2026-03-14 17:20:56","commit":"dca59a8afbcb9d852e17eb65eee52a55f78613d0","job_name":"PR-check","job_id":23090208497,"branch":"main"},"1773495254000000":{"status":"passed","date":"2026-03-14 13:34:14","commit":"12f3fd4278f6d53827e082b26c724dc9e13c7788","job_name":"Postcommit_asan","job_id":23087154243,"branch":"main"},"1773495175000000":{"status":"passed","date":"2026-03-14 13:32:55","commit":"7b7f59c71f27e2326a7776d6dd5863d41a005c73","job_name":"PR-check","job_id":23087448409,"branch":"main"},"1773494895000000":{"status":"passed","date":"2026-03-14 13:28:15","commit":"b2807c11a7982d173237b9ed80f477bc84c99a28","job_name":"Postcommit_asan","job_id":23087096661,"branch":"main"},"1773489674000000":{"status":"passed","date":"2026-03-14 12:01:14","commit":"a52984f48ec36fd8bb137a74541e2e74f8f54012","job_name":"Postcommit_asan","job_id":23085677696,"branch":"main"},"1773488352000000":{"status":"passed","date":"2026-03-14 11:39:12","commit":"d1f486c1ab24f785eb39e4a6347f3d0949b759b0","job_name":"PR-check","job_id":23076432021,"branch":"main"},"1773488108000000":{"status":"passed","date":"2026-03-14 11:35:08","commit":"d774235e49af4da77f35f387c5defaf1798642e7","job_name":"PR-check","job_id":23079817995,"branch":"main"},"1773485812000000":{"status":"passed","date":"2026-03-14 10:56:52","commit":"bc4a40d56fd5c0ce2f7d549044e9d2ef6214b7f0","job_name":"Postcommit_asan","job_id":23070611144,"branch":"main"},"1773485425000000":{"status":"passed","date":"2026-03-14 10:50:25","commit":"f5193a3a32959b822559586d4a9aef4d9ed16e4a","job_name":"PR-check","job_id":23081675479,"branch":"main"},"1773485307000000":{"status":"passed","date":"2026-03-14 10:48:27","commit":"476f8889f52d729a16097b545cde498236be9ab9","job_name":"PR-check","job_id":23068023330,"branch":"main"},"1773485102000000":{"status":"failure","date":"2026-03-14 10:45:02","commit":"a2ce07cbe86052674aca1965bfb397f46f5cf41f","job_name":"PR-check","job_id":23076607985,"branch":"main"},"1773481064000000":{"status":"passed","date":"2026-03-14 09:37:44","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23084133632,"branch":"main"},"1773463643000000":{"status":"passed","date":"2026-03-14 04:47:23","commit":"14c41b5dab0037e0ab2bb04c0787b16511294097","job_name":"Regression-run_Small_and_Medium","job_id":23074991149,"branch":"main"},"1773428652000000":{"status":"passed","date":"2026-03-13 19:04:12","commit":"d3b45dfa22bc3090e9c77516e0cc2ecde82a93df","job_name":"PR-check","job_id":23062317647,"branch":"main"},"1773427126000000":{"status":"passed","date":"2026-03-13 18:38:46","commit":"fc6691794d0e2f94a92b34914725d3831fee2535","job_name":"PR-check","job_id":23060898209,"branch":"main"},"1773423092000000":{"status":"passed","date":"2026-03-13 17:31:32","commit":"14b1052ed0267dcec8ff0492ed029ba3b52caac8","job_name":"PR-check","job_id":23059041434,"branch":"main"},"1773422300000000":{"status":"passed","date":"2026-03-13 17:18:20","commit":"ec157d5699d81f4b62ec9f69db9fe6e1ef1d6ef7","job_name":"PR-check","job_id":23056186233,"branch":"main"},"1773421818000000":{"status":"passed","date":"2026-03-13 17:10:18","commit":"b965f1f389f46357467ac73199b14034a5e36a1c","job_name":"Postcommit_asan","job_id":23060758486,"branch":"main"},"1773421534000000":{"status":"passed","date":"2026-03-13 17:05:34","commit":"d449a8cabb7c77c4e9c19deeeff18a840ecaf883","job_name":"PR-check","job_id":23056818501,"branch":"main"},"1773419832000000":{"status":"passed","date":"2026-03-13 16:37:12","commit":"bb4d182454be9ce2e910148a840322daa6fa9156","job_name":"PR-check","job_id":23054889866,"branch":"main"},"1773419396000000":{"status":"failure","date":"2026-03-13 16:29:56","commit":"69533a32cd75070c246d6abb4d121a3434fd29ce","job_name":"PR-check","job_id":23055881596,"branch":"main"},"1773419081000000":{"status":"passed","date":"2026-03-13 16:24:41","commit":"e92b8bbf8ca9883b328f20c01f448ba597015d73","job_name":"PR-check","job_id":23055606713,"branch":"main"},"1773418484000000":{"status":"passed","date":"2026-03-13 16:14:44","commit":"77e3edc6bda8c2635b2b614d21276adefaf7bf5b","job_name":"PR-check","job_id":23054890856,"branch":"main"},"1773418466000000":{"status":"passed","date":"2026-03-13 16:14:26","commit":"c3a3cf6a2398f62d592bb30a699a1c7498251787","job_name":"Postcommit_asan","job_id":23054570230,"branch":"main"},"1773418209000000":{"status":"failure","date":"2026-03-13 16:10:09","commit":"358fa229dc8ee382d0bc6060d10d5a65e238768d","job_name":"PR-check","job_id":23054476885,"branch":"main"},"1773417840000000":{"status":"passed","date":"2026-03-13 16:04:00","commit":"af1ab737ac34081d29f38bbc71bf21fa1702f6fd","job_name":"Postcommit_asan","job_id":23053099362,"branch":"main"},"1773417667000000":{"status":"passed","date":"2026-03-13 16:01:07","commit":"c96c60198bfa59406fd94b9490a5f8ed0ab5795d","job_name":"Postcommit_asan","job_id":23052790242,"branch":"main"},"1773417296000000":{"status":"passed","date":"2026-03-13 15:54:56","commit":"a24612ad59860bf811e49a5ca15884f0050d8bcf","job_name":"PR-check","job_id":23054186274,"branch":"main"},"1773417175000000":{"status":"passed","date":"2026-03-13 15:52:55","commit":"38455ed3bff6b199af6b13621f6a8fe559af8522","job_name":"PR-check","job_id":23052797685,"branch":"main"},"1773417033000000":{"status":"passed","date":"2026-03-13 15:50:33","commit":"c31f04ee3cb3a9972a19f4ee8b0848bf2180ecde","job_name":"PR-check","job_id":23054621490,"branch":"main"},"1773416973000000":{"status":"failure","date":"2026-03-13 15:49:33","commit":"7678236e573256b74fa6c4799caeb76b09645684","job_name":"Postcommit_asan","job_id":23053967928,"branch":"main"},"1773416913000000":{"status":"passed","date":"2026-03-13 15:48:33","commit":"31bc092ee1ef942c180de41c2c2ff2d23c764406","job_name":"Postcommit_asan","job_id":23054118246,"branch":"main"},"1773416723000000":{"status":"passed","date":"2026-03-13 15:45:23","commit":"106de92fce72898e4a9103c1cfbe4d9cd59e1cc2","job_name":"PR-check","job_id":23054146245,"branch":"main"},"1773416600000000":{"status":"failure","date":"2026-03-13 15:43:20","commit":"0d39a0ce2946a1bb38472000a14334277883785d","job_name":"PR-check","job_id":23052855885,"branch":"main"},"1773416491000000":{"status":"passed","date":"2026-03-13 15:41:31","commit":"30acea6972011a34d99f41ce1258b90dc65c3b47","job_name":"Postcommit_asan","job_id":23052415579,"branch":"main"},"1773414132000000":{"status":"passed","date":"2026-03-13 15:02:12","commit":"941eac0f1eeddec3cc0d7770b7f8fe6c584e5b27","job_name":"PR-check","job_id":23050671850,"branch":"main"},"1773413019000000":{"status":"passed","date":"2026-03-13 14:43:39","commit":"76995ddec63585100bdc68fc8d363b2f25cf1e14","job_name":"PR-check","job_id":23050126474,"branch":"main"},"1773412901000000":{"status":"failure","date":"2026-03-13 14:41:41","commit":"893c3b23679e2b451a117e483156b31c281cb74b","job_name":"PR-check","job_id":23050332572,"branch":"main"},"1773412877000000":{"status":"passed","date":"2026-03-13 14:41:17","commit":"814c7b3f4d7fe3f45d61f623a3ff3a7ffb0870c0","job_name":"Postcommit_asan","job_id":23050053326,"branch":"main"},"1773412278000000":{"status":"passed","date":"2026-03-13 14:31:18","commit":"dba91affc8b97e4fa1b64a129d7f0720ef44a567","job_name":"PR-check","job_id":23049775921,"branch":"main"},"1773409896000000":{"status":"passed","date":"2026-03-13 13:51:36","commit":"ee1e08cf41ccc5407e28457ff3dc38bbd54eb274","job_name":"PR-check","job_id":23049928982,"branch":"main"},"1773408162000000":{"status":"passed","date":"2026-03-13 13:22:42","commit":"ef23eca35059a170b13179b0d61a1fa100135107","job_name":"PR-check","job_id":23049040029,"branch":"main"},"1773407266000000":{"status":"passed","date":"2026-03-13 13:07:46","commit":"561a8dfd53e3020ecfd41dcad31807214b8b012d","job_name":"PR-check","job_id":23047585810,"branch":"main"},"1773407147000000":{"status":"passed","date":"2026-03-13 13:05:47","commit":"04c79bf03d87202cfaf904782bfc167e8ac1d5da","job_name":"PR-check","job_id":23048553198,"branch":"main"},"1773406369000000":{"status":"passed","date":"2026-03-13 12:52:49","commit":"1b012effdc52f8b1c1622dc9add30bfded1c7edf","job_name":"Postcommit_asan","job_id":23047927166,"branch":"main"},"1773405522000000":{"status":"passed","date":"2026-03-13 12:38:42","commit":"668f76c3e0c1649e020f3b49176c99ccfe6f8aac","job_name":"PR-check","job_id":23047028841,"branch":"main"},"1773404601000000":{"status":"passed","date":"2026-03-13 12:23:21","commit":"b1630af2d6d54f4ad460d19becfbb029e7a6e44b","job_name":"PR-check","job_id":23045877632,"branch":"main"},"1773402640000000":{"status":"passed","date":"2026-03-13 11:50:40","commit":"e579d2cdca84d3fa4864cb75f40f895acdd51d8b","job_name":"Postcommit_asan","job_id":23044877655,"branch":"main"},"1773401100000000":{"status":"passed","date":"2026-03-13 11:25:00","commit":"f38c970ef2df972ae69df654d2ddc026f946eac7","job_name":"PR-check","job_id":23043913593,"branch":"main"},"1773398367000000":{"status":"passed","date":"2026-03-13 10:39:27","commit":"fc99c848253eba018418c0e8d09455f2ba66844a","job_name":"PR-check","job_id":23042284918,"branch":"main"},"1773397572000000":{"status":"passed","date":"2026-03-13 10:26:12","commit":"f728ec9198b12beab4d506cacb0acb6c6892e83b","job_name":"PR-check","job_id":23043148469,"branch":"main"},"1773397083000000":{"status":"passed","date":"2026-03-13 10:18:03","commit":"9526f0fec62fd21a5b862b1f43a31021116badba","job_name":"PR-check","job_id":23041812735,"branch":"main"},"1773394916000000":{"status":"passed","date":"2026-03-13 09:41:56","commit":"9e64063a0513810d2ca042d191ba54714b974932","job_name":"Postcommit_asan","job_id":23042032235,"branch":"main"},"1773394033000000":{"status":"passed","date":"2026-03-13 09:27:13","commit":"18443ebafb7706487c1fa74616b793000248b315","job_name":"Postcommit_asan","job_id":23041580612,"branch":"main"},"1773392201000000":{"status":"passed","date":"2026-03-13 08:56:41","commit":"f113012ebd1dfcc030c28962ad99cb66091e3bac","job_name":"PR-check","job_id":23040735515,"branch":"main"},"1773391749000000":{"status":"passed","date":"2026-03-13 08:49:09","commit":"5c6c1dc2737edbc1457e8833e0002bf335fae2d3","job_name":"PR-check","job_id":23040231310,"branch":"main"},"1773391647000000":{"status":"passed","date":"2026-03-13 08:47:27","commit":"9ba885648b6ffb81fb7d4255d5cb45f50d44c1f7","job_name":"PR-check","job_id":23038755579,"branch":"main"},"1773389491000000":{"status":"passed","date":"2026-03-13 08:11:31","commit":"76160507b4bfb6e75f75bbe70881d72649590a0e","job_name":"Postcommit_asan","job_id":23040552823,"branch":"main"},"1773389368000000":{"status":"passed","date":"2026-03-13 08:09:28","commit":"ca10b96c9c6867368e798ebc8645e77fad9ca0f7","job_name":"PR-check","job_id":23039196379,"branch":"main"},"1773389248000000":{"status":"passed","date":"2026-03-13 08:07:28","commit":"5cb5c2782c8d321fae0fe13dc51bbbd04be737d3","job_name":"PR-check","job_id":23038548842,"branch":"main"},"1773389101000000":{"status":"passed","date":"2026-03-13 08:05:01","commit":"84b8dda8f4e4baef782ea555b0adb3882f8a89f8","job_name":"PR-check","job_id":23039037990,"branch":"main"},"1773388600000000":{"status":"passed","date":"2026-03-13 07:56:40","commit":"2337f9edf2644246da060ef99b17287514c996dc","job_name":"Postcommit_asan","job_id":23039881540,"branch":"main"},"1773377371000000":{"status":"passed","date":"2026-03-13 04:49:31","commit":"957efa2caf38f8689982f42c823b35735387987f","job_name":"Regression-run_Small_and_Medium","job_id":23029148017,"branch":"main"},"1773376892000000":{"status":"passed","date":"2026-03-13 04:41:32","commit":"d57159d2a332f8f6299a45b815a6ef646e09e196","job_name":"PR-check","job_id":23033279897,"branch":"main"},"1773374483000000":{"status":"passed","date":"2026-03-13 04:01:23","commit":"2d95fa6169be0afc4c99ce6be7947004906c5610","job_name":"PR-check","job_id":23031194054,"branch":"main"},"1773366191000000":{"status":"passed","date":"2026-03-13 01:43:11","commit":"b1b5f4ee910c14a3f139a03918feb23fba48bdee","job_name":"PR-check","job_id":23028965843,"branch":"main"},"1773365943000000":{"status":"passed","date":"2026-03-13 01:39:03","commit":"cb4a75c2c58b5c04938e70aa25d8194416b49431","job_name":"PR-check","job_id":23029069932,"branch":"main"},"1773364460000000":{"status":"passed","date":"2026-03-13 01:14:20","commit":"1319ee78cbe477a9ad4e0791f0f9280be9266814","job_name":"PR-check","job_id":23027847484,"branch":"main"},"1773361816000000":{"status":"passed","date":"2026-03-13 00:30:16","commit":"286a0829af8e29e0c9b30dc97715058700bb4c17","job_name":"PR-check","job_id":23026313832,"branch":"main"}}},"history_descriptions":{"ydb/core/statistics/service/ut/ColumnStatistics.EqWidthHistogram":{"1773708708000000":"assertion failed at ydb/core/statistics/service/ut/ut_column_statistics.cpp:206, virtual void NKikimr::NStat::NTestSuiteColumnStatistics::TTestCaseEqWidthHistogram::Execute_(NUnitTest::TTestContext &): (!responses.at(0).Success) \nTBackTrace::Capture() at /-S/util/system/backtrace.cpp:284:9\nGetCurrentTest at /-S/library/cpp/testing/unittest/registar.cpp:70:12\nUnRef at /-S/util/generic/ptr.h:637:13\noperator() at /-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:0:1\nTColoredProcessor::Run(std::__y1::function<void ()>, TBasicString<char, std::__y1::char_traits<char>> const&, char const*, bool) at /-S/library/cpp/testing/unittest/utmain.cpp:527:13\n~__value_func at /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:403:16\nUnRef at /-S/util/generic/ptr.h:637:13\nNUnitTest::TTestFactory::Execute() at /-S/library/cpp/testing/unittest/registar.cpp:0:19\nNUnitTest::RunMain(int, char**) at /-S/library/cpp/testing/unittest/utmain.cpp:0:44\n?? at ??:0:0\n?? at ??:0:0\n_start at ??:0:0\n","1773682225000000":"assertion failed at ydb/core/statistics/service/ut/ut_column_statistics.cpp:206, virtual void NKikimr::NStat::NTestSuiteColumnStatistics::TTestCaseEqWidthHistogram::Execute_(NUnitTest::TTestContext &): (!responses.at(0).Success) \nTBackTrace::Capture() at /-S/util/system/backtrace.cpp:284:9\nGetCurrentTest at /-S/library/cpp/testing/unittest/registar.cpp:70:12\nUnRef at /-S/util/generic/ptr.h:637:13\noperator() at /-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:0:1\nTColoredProcessor::Run(std::__y1::function<void ()>, TBasicString<char, std::__y1::char_traits<char>> const&, char const*, bool) at /-S/library/cpp/testing/unittest/utmain.cpp:527:13\n~__value_func at /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:403:16\nUnRef at /-S/util/generic/ptr.h:637:13\nNUnitTest::TTestFactory::Execute() at /-S/library/cpp/testing/unittest/registar.cpp:0:19\nNUnitTest::RunMain(int, char**) at /-S/library/cpp/testing/unittest/utmain.cpp:0:44\n?? at ??:0:0\n?? at ??:0:0\n_start at ??:0:0\n","1773678541000000":"Test crashed (return code: 100)\n==313543==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x00003bfae85e bp 0x7ffd2b5afc60 sp 0x7ffd2b5afb90 T0)\n==313543==The signal is caused by a WRITE memory access.\n==313543==Hint: address points to the zero page.\n    [[c:red]]#0 [[c:light-grey]]0x00003bfae85e in AtomicAdd [[c:light-grey]]/-S/library/cpp/deprecated/atomic/atomic_gcc.h:[[c:magenta]]51:12\n    [[c:red]]#1 [[c:light-grey]]0x00003bfae85e in Add [[c:light-grey]]/-S/library/cpp/monlib/counters/counters.h:[[c:magenta]]78:20\n    [[c:red]]#2 [[c:light-grey]]0x00003bfae85e in operator+[[c:green]]= [[c:light-grey]]/-S/library/cpp/monlib/counters/counters.h:[[c:magenta]]100:13\n    [[c:red]]#3 [[c:light-grey]]0x00003bfae85e in Delta [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/common/memusage.h:[[c:magenta]]36:22\n    [[c:red]]#4 [[c:light-grey]]0x00003bfae85e in Subtract [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/common/memusage.h:[[c:magenta]]32:13\n    [[c:red]]#5 [[c:light-grey]]0x00003bfae85e in ~TMemoryConsumerWithDropOnDestroy [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/common/memusage.h:[[c:magenta]]56:30\n    [[c:red]]#6 [[c:light-grey]]0x00003bfae85e in reset [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/optional:[[c:magenta]]319:15\n    [[c:red]]#7 [[c:light-grey]]0x00003bfae85e in TUnreplicatedBlobRecords [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:[[c:magenta]]186:32\n    [[c:red]]#8 [[c:light-grey]]0x00003bfae85e in [[c:cyan]]NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:[[c:magenta]]1083:15\n    [[c:red]]#9 [[c:light-grey]]0x00003bfadea7 in [[c:cyan]]NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:[[c:magenta]]1120:20\n    [[c:red]]#10 [[c:light-grey]]0x00003bf9c669 in [[c:cyan]]NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:[[c:magenta]]536:39\n    [[c:red]]#11 [[c:light-grey]]0x00003bf878ee in [[c:cyan]]NKikimr::TReplScheduler::StartReplication() [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:[[c:magenta]]304:13\n    [[c:red]]#12 [[c:light-grey]]0x00003bf84f66 in [[c:cyan]]NKikimr::TReplScheduler::Bootstrap() [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:[[c:magenta]]260:17\n    [[c:red]]#13 [[c:light-grey]]0x00003bf82e8e in NActors::TActorBootstrapped<NKikimr::TReplScheduler>[[c:cyan]]::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) [[c:light-grey]]/-S/ydb/library/actors/core/actor_bootstrapped.h:[[c:magenta]]26:22\n    [[c:red]]#14 [[c:light-grey]]0x00001d201007 in [[c:cyan]]NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) [[c:light-grey]]/-S/ydb/library/actors/core/actor.cpp:[[c:magenta]]354:17\n    [[c:red]]#15 [[c:light-grey]]0x0000396e5ca7 in [[c:cyan]]NActors::TTestActorRuntimeBase::SendInternal(TAutoPtr<NActors::IEventHandle, TDelete>, unsigned int, bool) [[c:light-grey]]/-S/ydb/library/actors/testlib/test_runtime.cpp:[[c:magenta]]1722:33\n    [[c:red]]#16 [[c:light-grey]]0x0000396de2c2 in [[c:cyan]]NActors::TTestActorRuntimeBase::DispatchEventsInternal(NActors::TDispatchOptions const&, TInstant) [[c:light-grey]]/-S/ydb/library/actors/testlib/test_runtime.cpp:[[c:magenta]]1311:45\n    [[c:red]]#17 [[c:light-grey]]0x0000396daee0 in DispatchEvents [[c:light-grey]]/-S/ydb/library/actors/testlib/test_runtime.cpp:[[c:magenta]]1107:16\n    [[c:red]]#18 [[c:light-grey]]0x0000396daee0 in [[c:cyan]]NActors::TTestActorRuntimeBase::DispatchEvents(NActors::TDispatchOptions const&) [[c:light-grey]]/-S/ydb/library/actors/testlib/test_runtime.cpp:[[c:magenta]]1098:16\n    [[c:red]]#19 [[c:light-grey]]0x000039b9c23f in [[c:cyan]]NKikimr::SetupBasicServices(NActors::TTestActorRuntime&, NKikimr::TAppPrepare&, bool, NKikimr::NFake::INode*, NKikimr::NFake::TStorage, NKikimrSharedCache::TSharedCacheConfig const*, bool, TVector<TIntrusivePtr<NKikimr::NFake::TProxyDS, TDefaultIntrusivePtrOps<NKikimr::NFake::TProxyDS>>, std::__y1::allocator<TIntrusivePtr<NKikimr::NFake::TProxyDS, TDefaultIntrusivePtrOps<NKikimr::NFake::TProxyDS>>>>) [[c:light-grey]]/-S\n..[snippet truncated]..\n() [[c:light-grey]]/-S/ydb/core/testlib/test_client.cpp:[[c:magenta]]611:9\n    [[c:red]]#22 [[c:light-grey]]0x00003ed50a5b in [[c:cyan]]NKikimr::Tests::TServer::TServer(TIntrusiveConstPtr<NKikimr::Tests::TServerSettings, TDefaultIntrusivePtrOps<NKikimr::Tests::TServerSettings>>, bool) [[c:light-grey]]/-S/ydb/core/testlib/test_client.cpp:[[c:magenta]]467:13\n    [[c:red]]#23 [[c:light-grey]]0x00003ed539fb in [[c:cyan]]NKikimr::Tests::TServer::TServer(NKikimr::Tests::TServerSettings const&, bool) [[c:light-grey]]/-S/ydb/core/testlib/test_client.cpp:[[c:magenta]]472:11\n    [[c:red]]#24 [[c:light-grey]]0x00004f5bc6eb in [[c:cyan]]NKikimr::NStat::TTestEnv::TTestEnv(unsigned int, unsigned int, bool, std::__y1::function<void (NKikimr::Tests::TServerSettings&)>) [[c:light-grey]]/-S/ydb/core/statistics/ut_common/ut_common.cpp:[[c:magenta]]55:18\n    [[c:red]]#25 [[c:light-grey]]0x00001b927c1b in [[c:cyan]]NKikimr::NStat::(anonymous namespace)::PrepareAndGetStatistics(std::__y1::vector<NKikimr::NStat::TColumnDesc, std::__y1::allocator<NKikimr::NStat::TColumnDesc>> const&, NKikimr::NStat::EStatType, std::__y1::vector<std::__y1::basic_string<char, std::__y1::char_traits<char>, std::__y1::allocator<char>>, std::__y1::allocator<std::__y1::basic_string<char, std::__y1::char_traits<char>, std::__y1::allocator<char>>>> const&) [[c:light-grey]]/-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:[[c:magenta]]91:14\n    [[c:red]]#26 [[c:light-grey]]0x00001b92bbea in [[c:cyan]]NKikimr::NStat::NTestSuiteColumnStatistics::TTestCaseEqWidthHistogram::Execute_(NUnitTest::TTestContext&) [[c:light-grey]]/-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:[[c:magenta]]196:26\n    [[c:red]]#27 [[c:light-grey]]0x00001b9436e7 in [[c:cyan]]operator() [[c:light-grey]]/-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:[[c:magenta]]124:1\n    [[c:red]]#28 [[c:light-grey]]0x00001b9436e7 in __invoke<(lambda at /-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:124:1) &> [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:[[c:magenta]]179:25\n    [[c:red]]#29 [[c:light-grey]]0x00001b9436e7 in __call<(lambda at /-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:124:1) &> [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:[[c:magenta]]251:5\n    [[c:red]]#30 [[c:light-grey]]0x00001b9436e7 in __invoke_r<void, (lambda at /-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:124:1) &> [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:[[c:magenta]]273:10\n    [[c:red]]#31 [[c:light-grey]]0x00001b9436e7 in [[c:cyan]]operator() [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:[[c:magenta]]167:12\n    [[c:red]]#32 [[c:light-grey]]0x00001b9436e7 in std::__y1::__function::__func<[[c:cyan]]NKikimr::NStat::NTestSuiteColumnStatistics::TCurrentTest::Execute()::'lambda'(), std::__y1::allocator<NKikimr::NStat::NTestSuiteColumnStatistics::TCurrentTest::Execute()::'lambda'()>, void ()>::operator()() [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:[[c:magenta]]319:10\n    [[c:red]]#33 [[c:light-grey]]0x00001c24e3e9 in [[c:cyan]]operator() [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:[[c:magenta]]436:12\n    [[c:red]]#34 [[c:light-grey]]0x00001c24e3e9 in [[c:cyan]]operator() [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:[[c:magenta]]995:10\n    [[c:red]]#35 [[c:light-grey]]0x00001c24e3e9 in [[c:cyan]]TColoredProcessor::Run(std::__y1::function<void ()>, TBasicString<char, std::__y1::char_traits<char>> const&, char const*, bool) [[c:light-grey]]/-S/library/cpp/testing/unittest/utmain.cpp:[[c:magenta]]527:20\n    [[c:red]]#36 [[c:light-grey]]0x00001c226ec7 in [[c:cyan]]NUnitTest::TTestBase::Run(std::__y1::function<void ()>, TBasicString<char, std::__y1::char_traits<char>> const&, char const*, bool) [[c:light-grey]]/-S/library/cpp/testing/unittest/registar.cpp:[[c:magenta]]378:18\n    [[c:red]]#37 [[c:light-grey]]0x00001b942a15 in [[c:cyan]]NKikimr::NStat::NTestSuiteColumnStatistics::TCurrentTest::Execute() [[c:light-grey]]/-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:[[c:magenta]]124:1\n    [[c:red]]#38 [[c:light-grey]]0x00001c22867f in [[c:cyan]]NUnitTest::TTestFactory::Execute() [[c:light-grey]]/-S/library/cpp/testing/unittest/registar.cpp:[[c:magenta]]499:19\n    [[c:red]]#39 [[c:light-grey]]0x00001c2484fc in [[c:cyan]]NUnitTest::RunMain(int, char**) [[c:light-grey]]/-S/library/cpp/testing/unittest/utmain.cpp:[[c:magenta]]899:44\n    [[c:red]]#40 [[c:light-grey]]0x7f57fee20d8f  (/lib/x86_64-linux-gnu/libc.so.6+0x29d8f) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\n    [[c:red]]#41 [[c:light-grey]]0x7f57fee20e3f in [[c:cyan]]__libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x29e3f) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\n    [[c:red]]#42 [[c:light-grey]]0x00001924a028 in [[c:cyan]]_start (/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/core/statistics/service/ut/ydb-core-statistics-service-ut+[[c:light-grey]]0x1924a028) (BuildId: 02c85e084f047a4f49053a1aac97fd72b1f3a955)\nSUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\n==313543==ABORTING","1773417175000000":"Test crashed (return code: -6)\nSee logs for more info"},"ydb/core/statistics/service/ut/ColumnStatistics.ManyColumns":{"1773735610000000":"Killed by timeout (600 s)","1773729526000000":"Killed by timeout (600 s)","1773722559000000":"Killed by timeout (600 s)","1773720290000000":"Killed by timeout (600 s)","1773719674000000":"Killed by timeout (600 s)","1773717082000000":"Killed by timeout (600 s)","1773708708000000":"Killed by timeout (600 s)","1773702149000000":"Killed by timeout (600 s)","1773702129000000":"Killed by timeout (600 s)","1773698398000000":"Killed by timeout (600 s)","1773694511000000":"Killed by timeout (600 s)","1773692039000000":"Killed by timeout (600 s)","1773690853000000":"Killed by timeout (600 s)","1773690441000000":"Killed by timeout (600 s)","1773690098000000":"Killed by timeout (600 s)","1773689559000000":"Killed by timeout (600 s)","1773689025000000":"Killed by timeout (600 s)","1773688476000000":"Killed by timeout (600 s)","1773687448000000":"Killed by timeout (600 s)","1773687172000000":"Killed by timeout (600 s)","1773687167000000":"Killed by timeout (600 s)","1773687158000000":"Killed by timeout (600 s)","1773687146000000":"Killed by timeout (600 s)","1773685622000000":"Killed by timeout (600 s)","1773685521000000":"Killed by timeout (600 s)","1773683098000000":"Killed by timeout (600 s)","1773682225000000":"Killed by timeout (600 s)","1773682040000000":"Killed by timeout (600 s)","1773681818000000":"Killed by timeout (600 s)","1773680250000000":"Killed by timeout (600 s)","1773679933000000":"Killed by timeout (600 s)","1773679449000000":"Killed by timeout (600 s)","1773679120000000":"Killed by timeout (600 s)","1773679007000000":"Killed by timeout (600 s)","1773678560000000":"Killed by timeout (600 s)","1773678541000000":"Test crashed (return code: 100)\n==314739==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x00003bfae85e bp 0x7ffd9360ba00 sp 0x7ffd9360b930 T0)\n==314739==The signal is caused by a WRITE memory access.\n==314739==Hint: address points to the zero page.\n    [[c:red]]#0 [[c:light-grey]]0x00003bfae85e in AtomicAdd [[c:light-grey]]/-S/library/cpp/deprecated/atomic/atomic_gcc.h:[[c:magenta]]51:12\n    [[c:red]]#1 [[c:light-grey]]0x00003bfae85e in Add [[c:light-grey]]/-S/library/cpp/monlib/counters/counters.h:[[c:magenta]]78:20\n    [[c:red]]#2 [[c:light-grey]]0x00003bfae85e in operator+[[c:green]]= [[c:light-grey]]/-S/library/cpp/monlib/counters/counters.h:[[c:magenta]]100:13\n    [[c:red]]#3 [[c:light-grey]]0x00003bfae85e in Delta [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/common/memusage.h:[[c:magenta]]36:22\n    [[c:red]]#4 [[c:light-grey]]0x00003bfae85e in Subtract [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/common/memusage.h:[[c:magenta]]32:13\n    [[c:red]]#5 [[c:light-grey]]0x00003bfae85e in ~TMemoryConsumerWithDropOnDestroy [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/common/memusage.h:[[c:magenta]]56:30\n    [[c:red]]#6 [[c:light-grey]]0x00003bfae85e in reset [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/optional:[[c:magenta]]319:15\n    [[c:red]]#7 [[c:light-grey]]0x00003bfae85e in TUnreplicatedBlobRecords [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:[[c:magenta]]186:32\n    [[c:red]]#8 [[c:light-grey]]0x00003bfae85e in [[c:cyan]]NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:[[c:magenta]]1083:15\n    [[c:red]]#9 [[c:light-grey]]0x00003bfadea7 in [[c:cyan]]NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:[[c:magenta]]1120:20\n    [[c:red]]#10 [[c:light-grey]]0x00003bf9c669 in [[c:cyan]]NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:[[c:magenta]]536:39\n    [[c:red]]#11 [[c:light-grey]]0x00003bf878ee in [[c:cyan]]NKikimr::TReplScheduler::StartReplication() [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:[[c:magenta]]304:13\n    [[c:red]]#12 [[c:light-grey]]0x00003bf84f66 in [[c:cyan]]NKikimr::TReplScheduler::Bootstrap() [[c:light-grey]]/-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:[[c:magenta]]260:17\n    [[c:red]]#13 [[c:light-grey]]0x00003bf82e8e in NActors::TActorBootstrapped<NKikimr::TReplScheduler>[[c:cyan]]::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) [[c:light-grey]]/-S/ydb/library/actors/core/actor_bootstrapped.h:[[c:magenta]]26:22\n    [[c:red]]#14 [[c:light-grey]]0x00001d201007 in [[c:cyan]]NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) [[c:light-grey]]/-S/ydb/library/actors/core/actor.cpp:[[c:magenta]]354:17\n    [[c:red]]#15 [[c:light-grey]]0x0000396e5ca7 in [[c:cyan]]NActors::TTestActorRuntimeBase::SendInternal(TAutoPtr<NActors::IEventHandle, TDelete>, unsigned int, bool) [[c:light-grey]]/-S/ydb/library/actors/testlib/test_runtime.cpp:[[c:magenta]]1722:33\n    [[c:red]]#16 [[c:light-grey]]0x0000396de2c2 in [[c:cyan]]NActors::TTestActorRuntimeBase::DispatchEventsInternal(NActors::TDispatchOptions const&, TInstant) [[c:light-grey]]/-S/ydb/library/actors/testlib/test_runtime.cpp:[[c:magenta]]1311:45\n    [[c:red]]#17 [[c:light-grey]]0x0000396daee0 in DispatchEvents [[c:light-grey]]/-S/ydb/library/actors/testlib/test_runtime.cpp:[[c:magenta]]1107:16\n    [[c:red]]#18 [[c:light-grey]]0x0000396daee0 in [[c:cyan]]NActors::TTestActorRuntimeBase::DispatchEvents(NActors::TDispatchOptions const&) [[c:light-grey]]/-S/ydb/library/actors/testlib/test_runtime.cpp:[[c:magenta]]1098:16\n    [[c:red]]#19 [[c:light-grey]]0x000039b9c23f in [[c:cyan]]NKikimr::SetupBasicServices(NActors::TTestActorRuntime&, NKikimr::TAppPrepare&, bool, NKikimr::NFake::INode*, NKikimr::NFake::TStorage, NKikimrSharedCache::TSharedCacheConfig const*, bool, TVector<TIntrusivePtr<NKikimr::NFake::TProxyDS, TDefaultIntrusivePtrOps<NKikimr::NFake::TProxyDS>>, std::__y1::allocator<TIntrusivePtr<NKikimr::NFake::TProxyDS, TDefaultIntrusivePtrOps<NKikimr::NFake::TProxyDS>>>>) [[c:light-grey]]/-S\n..[snippet truncated]..\nbf14 in [[c:cyan]]NKikimr::SetupTabletServices(NActors::TTestActorRuntime&, NKikimr::TAppPrepare*, bool, NKikimr::NFake::TStorage, NKikimrSharedCache::TSharedCacheConfig const*, bool, TVector<TIntrusivePtr<NKikimr::NFake::TProxyDS, TDefaultIntrusivePtrOps<NKikimr::NFake::TProxyDS>>, std::__y1::allocator<TIntrusivePtr<NKikimr::NFake::TProxyDS, TDefaultIntrusivePtrOps<NKikimr::NFake::TProxyDS>>>>) [[c:light-grey]]/-S/ydb/core/testlib/tablet_helpers.cpp:[[c:magenta]]652:9\n    [[c:red]]#21 [[c:light-grey]]0x00003ed52b93 in [[c:cyan]]NKikimr::Tests::TServer::Initialize() [[c:light-grey]]/-S/ydb/core/testlib/test_client.cpp:[[c:magenta]]611:9\n    [[c:red]]#22 [[c:light-grey]]0x00003ed50a5b in [[c:cyan]]NKikimr::Tests::TServer::TServer(TIntrusiveConstPtr<NKikimr::Tests::TServerSettings, TDefaultIntrusivePtrOps<NKikimr::Tests::TServerSettings>>, bool) [[c:light-grey]]/-S/ydb/core/testlib/test_client.cpp:[[c:magenta]]467:13\n    [[c:red]]#23 [[c:light-grey]]0x00003ed539fb in [[c:cyan]]NKikimr::Tests::TServer::TServer(NKikimr::Tests::TServerSettings const&, bool) [[c:light-grey]]/-S/ydb/core/testlib/test_client.cpp:[[c:magenta]]472:11\n    [[c:red]]#24 [[c:light-grey]]0x00004f5bc6eb in [[c:cyan]]NKikimr::NStat::TTestEnv::TTestEnv(unsigned int, unsigned int, bool, std::__y1::function<void (NKikimr::Tests::TServerSettings&)>) [[c:light-grey]]/-S/ydb/core/statistics/ut_common/ut_common.cpp:[[c:magenta]]55:18\n    [[c:red]]#25 [[c:light-grey]]0x00001b939a1e in [[c:cyan]]NKikimr::NStat::NTestSuiteColumnStatistics::TTestCaseManyColumns::Execute_(NUnitTest::TTestContext&) [[c:light-grey]]/-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:[[c:magenta]]299:18\n    [[c:red]]#26 [[c:light-grey]]0x00001b9436e7 in [[c:cyan]]operator() [[c:light-grey]]/-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:[[c:magenta]]124:1\n    [[c:red]]#27 [[c:light-grey]]0x00001b9436e7 in __invoke<(lambda at /-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:124:1) &> [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:[[c:magenta]]179:25\n    [[c:red]]#28 [[c:light-grey]]0x00001b9436e7 in __call<(lambda at /-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:124:1) &> [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:[[c:magenta]]251:5\n    [[c:red]]#29 [[c:light-grey]]0x00001b9436e7 in __invoke_r<void, (lambda at /-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:124:1) &> [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:[[c:magenta]]273:10\n    [[c:red]]#30 [[c:light-grey]]0x00001b9436e7 in [[c:cyan]]operator() [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:[[c:magenta]]167:12\n    [[c:red]]#31 [[c:light-grey]]0x00001b9436e7 in std::__y1::__function::__func<[[c:cyan]]NKikimr::NStat::NTestSuiteColumnStatistics::TCurrentTest::Execute()::'lambda'(), std::__y1::allocator<NKikimr::NStat::NTestSuiteColumnStatistics::TCurrentTest::Execute()::'lambda'()>, void ()>::operator()() [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:[[c:magenta]]319:10\n    [[c:red]]#32 [[c:light-grey]]0x00001c24e3e9 in [[c:cyan]]operator() [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:[[c:magenta]]436:12\n    [[c:red]]#33 [[c:light-grey]]0x00001c24e3e9 in [[c:cyan]]operator() [[c:light-grey]]/-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:[[c:magenta]]995:10\n    [[c:red]]#34 [[c:light-grey]]0x00001c24e3e9 in [[c:cyan]]TColoredProcessor::Run(std::__y1::function<void ()>, TBasicString<char, std::__y1::char_traits<char>> const&, char const*, bool) [[c:light-grey]]/-S/library/cpp/testing/unittest/utmain.cpp:[[c:magenta]]527:20\n    [[c:red]]#35 [[c:light-grey]]0x00001c226ec7 in [[c:cyan]]NUnitTest::TTestBase::Run(std::__y1::function<void ()>, TBasicString<char, std::__y1::char_traits<char>> const&, char const*, bool) [[c:light-grey]]/-S/library/cpp/testing/unittest/registar.cpp:[[c:magenta]]378:18\n    [[c:red]]#36 [[c:light-grey]]0x00001b942a15 in [[c:cyan]]NKikimr::NStat::NTestSuiteColumnStatistics::TCurrentTest::Execute() [[c:light-grey]]/-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:[[c:magenta]]124:1\n    [[c:red]]#37 [[c:light-grey]]0x00001c22867f in [[c:cyan]]NUnitTest::TTestFactory::Execute() [[c:light-grey]]/-S/library/cpp/testing/unittest/registar.cpp:[[c:magenta]]499:19\n    [[c:red]]#38 [[c:light-grey]]0x00001c2484fc in [[c:cyan]]NUnitTest::RunMain(int, char**) [[c:light-grey]]/-S/library/cpp/testing/unittest/utmain.cpp:[[c:magenta]]899:44\n    [[c:red]]#39 [[c:light-grey]]0x7fd319e3dd8f  (/lib/x86_64-linux-gnu/libc.so.6+0x29d8f) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\n    [[c:red]]#40 [[c:light-grey]]0x7fd319e3de3f in [[c:cyan]]__libc_start_main (/lib/x86_64-linux-gnu/libc.so.6+0x29e3f) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\n    [[c:red]]#41 [[c:light-grey]]0x00001924a028 in [[c:cyan]]_start (/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/core/statistics/service/ut/ydb-core-statistics-service-ut+[[c:light-grey]]0x1924a028) (BuildId: 02c85e084f047a4f49053a1aac97fd72b1f3a955)\nSUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\n==314739==ABORTING","1773678224000000":"Killed by timeout (600 s)","1773677952000000":"Killed by timeout (600 s)","1773677718000000":"Killed by timeout (600 s)","1773677510000000":"Killed by timeout (600 s)","1773677230000000":"Killed by timeout (600 s)","1773677073000000":"Killed by timeout (600 s)","1773676288000000":"Killed by timeout (600 s)","1773675427000000":"Killed by timeout (600 s)","1773675081000000":"Killed by timeout (600 s)","1773674966000000":"Killed by timeout (600 s)","1773674896000000":"Killed by timeout (600 s)","1773674709000000":"Killed by timeout (600 s)","1773674432000000":"Killed by timeout (600 s)","1773672978000000":"Killed by timeout (600 s)","1773672747000000":"Killed by timeout (600 s)","1773671187000000":"Killed by timeout (600 s)","1773668553000000":"Killed by timeout (600 s)","1773668400000000":"Killed by timeout (600 s)","1773667338000000":"Killed by timeout (600 s)","1773667038000000":"Killed by timeout (600 s)","1773666791000000":"Killed by timeout (600 s)","1773665446000000":"Killed by timeout (600 s)","1773663833000000":"Killed by timeout (600 s)","1773661068000000":"Killed by timeout (600 s)","1773660879000000":"Killed by timeout (600 s)","1773659987000000":"Killed by timeout (600 s)","1773659879000000":"Killed by timeout (600 s)","1773655409000000":"Killed by timeout (600 s)","1773653768000000":"Killed by timeout (600 s)","1773653670000000":"Killed by timeout (600 s)","1773651160000000":"Killed by timeout (600 s)","1773644268000000":"Killed by timeout (600 s)","1773643731000000":"Killed by timeout (600 s)","1773642105000000":"Killed by timeout (600 s)","1773637210000000":"Killed by timeout (600 s)","1773636520000000":"Killed by timeout (600 s)","1773634122000000":"Killed by timeout (600 s)","1773627098000000":"Killed by timeout (600 s)","1773626829000000":"Killed by timeout (600 s)","1773622583000000":"Killed by timeout (600 s)","1773617923000000":"Killed by timeout (600 s)","1773615765000000":"Killed by timeout (600 s)","1773615308000000":"Killed by timeout (600 s)","1773608284000000":"Killed by timeout (600 s)","1773606561000000":"Killed by timeout (600 s)","1773592886000000":"Killed by timeout (600 s)","1773572070000000":"Killed by timeout (600 s)","1773570977000000":"Killed by timeout (600 s)","1773562367000000":"Killed by timeout (600 s)","1773562210000000":"Killed by timeout (600 s)","1773549981000000":"Killed by timeout (600 s)","1773547892000000":"Killed by timeout (600 s)","1773526956000000":"Killed by timeout (600 s)","1773519197000000":"Killed by timeout (600 s)","1773517997000000":"Killed by timeout (600 s)","1773510184000000":"Killed by timeout (600 s)","1773508856000000":"Killed by timeout (600 s)","1773495254000000":"Killed by timeout (600 s)","1773495175000000":"Killed by timeout (600 s)","1773494895000000":"Killed by timeout (600 s)","1773489674000000":"Killed by timeout (600 s)","1773488352000000":"Killed by timeout (600 s)","1773488108000000":"Killed by timeout (600 s)","1773485812000000":"Killed by timeout (600 s)","1773485425000000":"Killed by timeout (600 s)","1773485307000000":"Killed by timeout (600 s)","1773485102000000":"Killed by timeout (600 s)","1773481064000000":"Killed by timeout (600 s)","1773463643000000":"Killed by timeout (600 s)","1773428652000000":"Killed by timeout (600 s)","1773427126000000":"Killed by timeout (600 s)","1773423092000000":"Killed by timeout (600 s)","1773422300000000":"Killed by timeout (600 s)","1773421818000000":"Killed by timeout (600 s)","1773421534000000":"Killed by timeout (600 s)","1773419832000000":"Killed by timeout (600 s)","1773419396000000":"Killed by timeout (600 s)","1773419081000000":"Killed by timeout (600 s)","1773418484000000":"Killed by timeout (600 s)","1773418466000000":"Killed by timeout (600 s)","1773418209000000":"Killed by timeout (600 s)","1773417840000000":"Killed by timeout (600 s)","1773417667000000":"Killed by timeout (600 s)","1773417296000000":"Killed by timeout (600 s)","1773417175000000":"Test crashed (return code: -6)\nSee logs for more info","1773417033000000":"Killed by timeout (600 s)","1773416973000000":"Killed by timeout (600 s)","1773416913000000":"Killed by timeout (600 s)","1773416723000000":"Killed by timeout (600 s)","1773416600000000":"Killed by timeout (600 s)","1773416491000000":"Killed by timeout (600 s)","1773414132000000":"Killed by timeout (600 s)","1773413019000000":"Killed by timeout (600 s)","1773412901000000":"Killed by timeout (600 s)","1773412877000000":"Killed by timeout (600 s)","1773412278000000":"Killed by timeout (600 s)","1773409896000000":"Killed by timeout (600 s)","1773408162000000":"Killed by timeout (600 s)","1773407266000000":"Killed by timeout (600 s)","1773407147000000":"Killed by timeout (600 s)","1773406369000000":"Killed by timeout (600 s)","1773405522000000":"Killed by timeout (600 s)","1773405071000000":"Killed by timeout (600 s)","1773404601000000":"Killed by timeout (600 s)","1773402640000000":"Killed by timeout (600 s)","1773401100000000":"Killed by timeout (600 s)","1773398367000000":"Killed by timeout (600 s)","1773397572000000":"Killed by timeout (600 s)","1773397083000000":"Killed by timeout (600 s)","1773394916000000":"Killed by timeout (600 s)","1773394033000000":"Killed by timeout (600 s)","1773392201000000":"Killed by timeout (600 s)","1773391749000000":"Killed by timeout (600 s)","1773391647000000":"Killed by timeout (600 s)","1773389491000000":"Killed by timeout (600 s)","1773389368000000":"Killed by timeout (600 s)","1773389248000000":"Killed by timeout (600 s)","1773389101000000":"Killed by timeout (600 s)","1773388600000000":"Killed by timeout (600 s)","1773377371000000":"Killed by timeout (600 s)","1773376892000000":"Killed by timeout (600 s)","1773374483000000":"Killed by timeout (600 s)","1773366191000000":"Killed by timeout (600 s)","1773365943000000":"Killed by timeout (600 s)","1773364460000000":"Killed by timeout (600 s)","1773361816000000":"Killed by timeout (600 s)"},"ydb/public/sdk/cpp/src/client/topic/ut/with_direct_read_ut/Describe.DescribePartitionPermissions":{"1773719674000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C21B2EB\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C70B5EB\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDE0650\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDE82F7\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDE82F7\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDE82F7\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDE82F7\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDE82F7\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDE82F7\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C739BA9\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C739BA9\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C739BA9\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C712687\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDE75D7\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C713E3F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C733CBC\n16. ??:0: ?? @ 0x7FCF6CA32D8F\n17. ??:0: ?? @ 0x7FCF6CA32E3F\n18. ??:0: ?? @ 0x19626028\n","1773698398000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C1F52FB\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6E55FB\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDBA650\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDC22F7\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDC22F7\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDC22F7\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDC22F7\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDC22F7\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDC22F7\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C713BB9\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C713BB9\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C713BB9\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6EC697\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDC15D7\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6EDE4F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C70DCCC\n16. ??:0: ?? @ 0x7F3E6D4BBD8F\n17. ??:0: ?? @ 0x7F3E6D4BBE3F\n18. ??:0: ?? @ 0x19600028\n","1773687172000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20A80B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6FAB0B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDCFB60\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDD7807\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD7807\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD7807\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD7807\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDD7807\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDD7807\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C7290C9\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C7290C9\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C7290C9\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C701BA7\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDD6AE7\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C70335F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C7231DC\n16. ??:0: ?? @ 0x7FEA3958ED8F\n17. ??:0: ?? @ 0x7FEA3958EE3F\n18. ??:0: ?? @ 0x19605028\n","1773672747000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20580B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F5B0B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDCAB60\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDD2807\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2807\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2807\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2807\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDD2807\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDD2807\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C7240C9\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C7240C9\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C7240C9\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6FCBA7\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDD1AE7\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6FE35F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C71E1DC\n16. ??:0: ?? @ 0x7F61BF784D8F\n17. ??:0: ?? @ 0x7F61BF784E3F\n18. ??:0: ?? @ 0x19600028\n","1773651160000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20580B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F5B0B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDCAB60\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDD2807\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2807\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2807\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2807\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDD2807\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDD2807\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C7240C9\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C7240C9\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C7240C9\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6FCBA7\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDD1AE7\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6FE35F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C71E1DC\n16. ??:0: ?? @ 0x7FDE674EAD8F\n17. ??:0: ?? @ 0x7FDE674EAE3F\n18. ??:0: ?? @ 0x19600028\n","1773634122000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20580B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F5B0B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDCAB60\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDD2807\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2807\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2807\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2807\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDD2807\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDD2807\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C7240C9\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C7240C9\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C7240C9\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6FCBA7\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDD1AE7\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6FE35F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C71E1DC\n16. ??:0: ?? @ 0x7FB8B30B3D8F\n17. ??:0: ?? @ 0x7FB8B30B3E3F\n18. ??:0: ?? @ 0x19600028\n","1773519197000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C260D1B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C75101B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BE26070\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BE2DD17\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BE2DD17\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BE2DD17\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BE2DD17\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BE2DD17\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BE2DD17\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C77F5D9\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C77F5D9\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C77F5D9\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C7580B7\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BE2CFF7\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C75986F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C7796EC\n16. ??:0: ?? @ 0x7FDBDBDDFD8F\n17. ??:0: ?? @ 0x7FDBDBDDFE3F\n18. ??:0: ?? @ 0x1962C028\n","1773517997000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20A76B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6FAA6B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDCFAC0\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDD7767\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD7767\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD7767\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD7767\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDD7767\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDD7767\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C729029\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C729029\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C729029\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C701B07\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDD6A47\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C7032BF\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C72313C\n16. ??:0: ?? @ 0x7F727CB97D8F\n17. ??:0: ?? @ 0x7F727CB97E3F\n18. ??:0: ?? @ 0x19605028\n","1773421818000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20576B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F5A6B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDCAAC0\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDD2767\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDD2767\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDD2767\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C724029\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C724029\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C724029\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6FCB07\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDD1A47\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6FE2BF\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C71E13C\n16. ??:0: ?? @ 0x7F1B9BA9DD8F\n17. ??:0: ?? @ 0x7F1B9BA9DE3F\n18. ??:0: ?? @ 0x19600028\n","1773418484000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20576B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F5A6B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDCAAC0\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDD2767\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDD2767\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDD2767\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C724029\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C724029\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C724029\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6FCB07\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDD1A47\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6FE2BF\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C71E13C\n16. ??:0: ?? @ 0x7F1B9BA9DD8F\n17. ??:0: ?? @ 0x7F1B9BA9DE3F\n18. ??:0: ?? @ 0x19600028\n","1773417667000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20576B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F5A6B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDCAAC0\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDD2767\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDD2767\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDD2767\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C724029\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C724029\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C724029\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6FCB07\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDD1A47\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6FE2BF\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C71E13C\n16. ??:0: ?? @ 0x7FE12DA60D8F\n17. ??:0: ?? @ 0x7FE12DA60E3F\n18. ??:0: ?? @ 0x19600028\n","1773416723000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20576B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F5A6B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDCAAC0\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDD2767\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDD2767\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDD2767\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDD2767\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C724029\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C724029\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C724029\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6FCB07\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDD1A47\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6FE2BF\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C71E13C\n16. ??:0: ?? @ 0x7F8EAF620D8F\n17. ??:0: ?? @ 0x7F8EAF620E3F\n18. ??:0: ?? @ 0x19600028\n","1773414132000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C22A77B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C71AA7B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDEFAC0\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDF7767\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDF7767\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDF7767\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDF7767\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDF7767\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDF7767\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C749039\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C749039\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C749039\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C721B17\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDF6A47\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C7232CF\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C74314C\n16. ??:0: ?? @ 0x7FCF8DC01D8F\n17. ??:0: ?? @ 0x7FCF8DC01E3F\n18. ??:0: ?? @ 0x19625028\n","1773409896000000":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C22277B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C712A7B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDE7AC0\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDEF767\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDEF767\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDEF767\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDEF767\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDEF767\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDEF767\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C741039\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C741039\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C741039\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C719B17\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDEEA47\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C71B2CF\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C73B14C\n16. ??:0: ?? @ 0x7F5EFACA5D8F\n17. ??:0: ?? @ 0x7F5EFACA5E3F\n18. ??:0: ?? @ 0x1961D028\n"},"ydb/core/kqp/ut/sysview/KqpSystemView.Sessions":{"1773708708000000":"assertion failed at ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:208, virtual void NKikimr::NKqp::NTestSuiteKqpSystemView::TTestCaseSessions::Execute_(NUnitTest::TTestContext &): (result.IsSuccess()) <main>: Fatal: Execution, code: 1060\n    <main>: Fatal: util/generic/maybe.cpp:5: TMaybe is empty, value type: TBasicString<char, std::char_traits<char>>, code: 1\n\n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C355D7B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C8519EB\n2. /tmp//-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:208: Execute_ @ 0x1BF0D74D\n3. /tmp//-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36: operator() @ 0x1BF76057\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36:1) &> @ 0x1BF76057\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36:1) &> @ 0x1BF76057\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36:1) &> @ 0x1BF76057\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BF76057\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BF76057\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C889599\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C889599\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C889599\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C858A87\n13. /tmp//-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36: Execute @ 0x1BF75385\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C85A23F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C8836AC\n16. ??:0: ?? @ 0x7F0AA3FD5D8F\n17. ??:0: ?? @ 0x7F0AA3FD5E3F\n18. ??:0: ?? @ 0x19442028\n","1773682225000000":"assertion failed at ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:208, virtual void NKikimr::NKqp::NTestSuiteKqpSystemView::TTestCaseSessions::Execute_(NUnitTest::TTestContext &): (result.IsSuccess()) <main>: Fatal: Execution, code: 1060\n    <main>: Fatal: util/generic/maybe.cpp:5: TMaybe is empty, value type: TBasicString<char, std::char_traits<char>>, code: 1\n\n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C355D5B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C8519CB\n2. /tmp//-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:208: Execute_ @ 0x1BF0D73D\n3. /tmp//-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36: operator() @ 0x1BF76037\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36:1) &> @ 0x1BF76037\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36:1) &> @ 0x1BF76037\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36:1) &> @ 0x1BF76037\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BF76037\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BF76037\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C889579\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C889579\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C889579\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C858A67\n13. /tmp//-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36: Execute @ 0x1BF75365\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C85A21F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C88368C\n16. ??:0: ?? @ 0x7F6F75552D8F\n17. ??:0: ?? @ 0x7F6F75552E3F\n18. ??:0: ?? @ 0x19442028\n"},"ydb/core/http_proxy/ut/inside_ydb_ut/TestKinesisHttpProxy.TestListStreamConsumersWithToken":{},"ydb/public/sdk/cpp/src/client/topic/ut/with_direct_read_ut/TxUsage.Sinks_Oltp_WriteToTopic_3_Table":{"1773562210000000":"assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/ut_utils/txusage_fixture.cpp:1022, void NYdb::Dev::NTopic::NTests::NTxUsage::TFixture::TestWriteToTopic26(): (messages.size() == 3) failed: (0 != 3) \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20880B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F8B0B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/ut_utils/txusage_fixture.cpp:1022: TestWriteToTopic26 @ 0x502A9B81\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6: operator() @ 0x1BE473A7\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE473A7\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE473A7\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE473A7\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BE473A7\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BE473A7\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C7270C9\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C7270C9\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C7270C9\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6FFBA7\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6: Execute @ 0x1BE466D5\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C70135F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C7211DC\n16. ??:0: ?? @ 0x7F9720D65D8F\n17. ??:0: ?? @ 0x7F9720D65E3F\n18. ??:0: ?? @ 0x19603028\n","1773547892000000":"assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/ut_utils/txusage_fixture.cpp:1022, void NYdb::Dev::NTopic::NTests::NTxUsage::TFixture::TestWriteToTopic26(): (messages.size() == 3) failed: (0 != 3) \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C20880B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F8B0B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/ut_utils/txusage_fixture.cpp:1022: TestWriteToTopic26 @ 0x502A9B81\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6: operator() @ 0x1BE473A7\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE473A7\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE473A7\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE473A7\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BE473A7\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BE473A7\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C7270C9\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C7270C9\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C7270C9\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6FFBA7\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6: Execute @ 0x1BE466D5\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C70135F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C7211DC\n16. ??:0: ?? @ 0x7F9720D65D8F\n17. ??:0: ?? @ 0x7F9720D65E3F\n18. ??:0: ?? @ 0x19603028\n","1773412278000000":"assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/ut_utils/txusage_fixture.cpp:1024, void NYdb::Dev::NTopic::NTests::NTxUsage::TFixture::TestWriteToTopic26(): (messages.size() == 3) failed: (1 != 3) \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C22577B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C715A7B\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/ut_utils/txusage_fixture.cpp:1024: TestWriteToTopic26 @ 0x5039C0F1\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6: operator() @ 0x1BE64317\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE64317\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE64317\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE64317\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BE64317\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BE64317\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C744039\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C744039\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C744039\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C71CB17\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6: Execute @ 0x1BE63645\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C71E2CF\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C73E14C\n16. ??:0: ?? @ 0x7F6173C70D8F\n17. ??:0: ?? @ 0x7F6173C70E3F\n18. ??:0: ?? @ 0x19620028\n"},"yql/essentials/core/cbo/cbo_optimizer_new.cpp.clang_format":{"1773708708000000":"--- L\n+++ R\n@@ -378,7 +378,7 @@\n \n     auto result = TOptimizerStatistics(outputType, newCard, newNCols, newByteSize, cost,\n                                        leftKeyColumns ? leftStats.KeyColumns : (rightKeyColumns ? rightStats.KeyColumns : TIntrusivePtr<TOptimizerStatistics::TKeyColumns>()));\n-    //result.Selectivity = std::pow(selectivity, 0.7);\n+    // result.Selectivity = std::pow(selectivity, 0.7);\n     result.Selectivity = selectivity;\n     return result;\n }","1773682225000000":"--- L\n+++ R\n@@ -378,7 +378,7 @@\n \n     auto result = TOptimizerStatistics(outputType, newCard, newNCols, newByteSize, cost,\n                                        leftKeyColumns ? leftStats.KeyColumns : (rightKeyColumns ? rightStats.KeyColumns : TIntrusivePtr<TOptimizerStatistics::TKeyColumns>()));\n-    //result.Selectivity = std::pow(selectivity, 0.7);\n+    // result.Selectivity = std::pow(selectivity, 0.7);\n     result.Selectivity = selectivity;\n     return result;\n }"},"ydb/tests/functional/statistics/py3test.[test_analyze.py 0/10] chunk":{"1773735610000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 595.66s\nKilled by timeout (600 s)","1773729526000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 584.14s\nKilled by timeout (600 s)","1773722559000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.05s","1773720290000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.15s","1773719674000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 609.77s","1773717082000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.94s","1773708708000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.79s","1773702149000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.24s","1773698398000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 595.10s\nKilled by timeout (600 s)","1773694511000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.51s","1773692039000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.68s","1773690853000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.93s","1773690441000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 592.86s\nKilled by timeout (600 s)","1773690098000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.48s","1773689559000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 613.33s","1773689025000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.10s","1773688476000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.44s","1773687172000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.81s","1773687167000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.77s","1773687158000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.24s","1773687146000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.67s","1773685622000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.17s","1773685521000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 593.63s\nKilled by timeout (600 s)","1773683098000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.55s","1773682225000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 589.04s\nKilled by timeout (600 s)","1773681818000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.98s","1773680250000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.24s","1773679933000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.00s","1773679449000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 623.62s","1773679120000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 588.62s\nKilled by timeout (600 s)","1773679007000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.91s","1773678560000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.64s","1773678224000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.52s","1773677952000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 613.92s","1773677718000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.15s","1773677510000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.55s","1773677230000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.38s","1773677073000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.10s","1773676288000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 619.61s","1773675427000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 619.37s","1773675081000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.43s","1773674966000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.42s","1773674896000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.01s","1773674709000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.77s","1773674432000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.08s","1773672978000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.25s","1773672747000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.73s","1773668553000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 595.32s\nKilled by timeout (600 s)","1773668400000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.27s","1773667338000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.25s","1773667038000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.99s","1773666791000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 589.09s\nKilled by timeout (600 s)","1773665446000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 613.97s","1773663833000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.48s","1773661068000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 595.32s\nKilled by timeout (600 s)","1773660879000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.48s","1773659987000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.82s","1773659879000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.16s","1773655409000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 618.04s","1773653768000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.86s","1773651160000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.03s","1773644268000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.13s","1773643731000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.87s","1773642105000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.94s","1773636520000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 613.88s","1773634122000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 617.07s","1773627098000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.37s","1773626829000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.99s","1773622583000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.17s","1773617923000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.17s","1773615765000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 596.08s\nKilled by timeout (600 s)","1773615308000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.10s","1773608284000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 618.51s","1773592886000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.07s","1773572070000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.92s","1773570977000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.07s","1773562367000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.42s","1773562210000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.53s","1773547892000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.53s","1773526956000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 587.78s\nKilled by timeout (600 s)","1773519197000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.67s","1773517997000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.17s","1773510184000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.16s","1773508856000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.63s","1773495175000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.24s","1773494895000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.25s","1773489674000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.29s","1773488352000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.59s","1773488108000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.25s","1773485812000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.36s","1773485425000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.27s","1773485307000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.42s","1773481064000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.88s","1773463643000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.09s","1773428652000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.69s","1773427126000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.42s","1773423092000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.88s","1773422300000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 619.58s","1773421818000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.20s","1773421534000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.01s","1773419832000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.85s","1773419396000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.37s","1773419081000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.36s","1773418484000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.73s","1773418466000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 619.32s","1773418209000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.06s","1773417840000000":"1 test: 1 - GOOD\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (good) duration: 595.33s\nKilled by timeout (600 s)","1773417667000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.62s","1773417296000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.29s","1773417033000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.54s","1773416913000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 618.97s","1773416491000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.21s","1773414132000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.73s","1773413019000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.73s","1773412901000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 612.87s","1773412877000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.88s","1773412278000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.39s","1773409896000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.50s","1773408162000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.34s","1773407266000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.33s","1773407147000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.96s","1773406369000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 618.69s","1773405522000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 625.30s","1773405071000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.57s","1773404601000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 613.50s","1773402640000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.67s","1773401100000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 625.99s","1773398367000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 616.09s","1773397572000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 621.63s","1773397083000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.16s","1773394916000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 613.86s","1773394033000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.58s","1773392201000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 619.97s","1773391749000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 613.51s","1773391647000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 619.39s","1773389368000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 628.22s","1773389248000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 620.59s","1773389101000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 622.04s","1773388600000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 617.80s","1773376892000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 619.14s","1773374483000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 617.80s","1773366191000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 624.42s","1773365943000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 615.79s","1773364460000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 618.92s","1773361816000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 618.24s"},"ydb/tests/datashard/truncate/concurrency/py3test.[test_truncate_table_concurrency.py 0/10] chunk":{"1773735610000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 45.79s\nKilled by timeout (60 s)","1773729526000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 46.10s\nKilled by timeout (60 s)","1773720290000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 52.65s\nKilled by timeout (60 s)","1773719674000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 46.19s\nKilled by timeout (60 s)","1773717082000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 62.36s","1773702149000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 50.68s\nKilled by timeout (60 s)","1773702129000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 53.84s\nKilled by timeout (60 s)","1773698398000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 50.68s\nKilled by timeout (60 s)","1773694511000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 48.03s\nKilled by timeout (60 s)","1773690853000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 49.57s\nKilled by timeout (60 s)","1773690098000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 41.57s\nKilled by timeout (60 s)","1773688476000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.35s\nKilled by timeout (60 s)","1773687448000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 46.26s\nKilled by timeout (60 s)","1773687172000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 54.92s\nKilled by timeout (60 s)","1773687167000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 49.01s\nKilled by timeout (60 s)","1773685622000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 44.90s\nKilled by timeout (60 s)","1773685521000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 42.77s\nKilled by timeout (60 s)","1773681818000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 51.12s\nKilled by timeout (60 s)","1773679933000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 55.45s\nKilled by timeout (60 s)","1773679449000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 53.23s\nKilled by timeout (60 s)","1773679120000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 69.07s","1773678560000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 84.81s","1773678541000000":"1 test: 1 - FAIL\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (fail) duration: 54.78s\nKilled by timeout (60 s)","1773677952000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 39.89s\nKilled by timeout (60 s)","1773677718000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 54.18s\nKilled by timeout (60 s)","1773677510000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 68.22s","1773677230000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 49.92s","1773677073000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 49.42s\nKilled by timeout (60 s)","1773676288000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 39.99s\nKilled by timeout (60 s)","1773675081000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 52.50s\nKilled by timeout (60 s)","1773674896000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 54.89s\nKilled by timeout (60 s)","1773674709000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 43.38s\nKilled by timeout (60 s)","1773674432000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 50.02s\nKilled by timeout (60 s)","1773672978000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 39.56s\nKilled by timeout (60 s)","1773672747000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 45.49s\nKilled by timeout (60 s)","1773671187000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 51.04s\nKilled by timeout (60 s)","1773668553000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.73s\nKilled by timeout (60 s)","1773667338000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 39.56s\nKilled by timeout (60 s)","1773667038000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 74.05s","1773666791000000":"1 test: 1 - FAIL\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (fail) duration: 50.54s\nKilled by timeout (60 s)","1773663833000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 50.07s\nKilled by timeout (60 s)","1773661068000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.73s\nKilled by timeout (60 s)","1773653768000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 44.21s\nKilled by timeout (60 s)","1773653670000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 80.56s","1773651160000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.36s\nKilled by timeout (60 s)","1773643731000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 44.36s\nKilled by timeout (60 s)","1773642105000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 48.12s\nKilled by timeout (60 s)","1773634122000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 43.00s\nKilled by timeout (60 s)","1773627098000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 40.57s\nKilled by timeout (60 s)","1773626829000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 52.32s\nKilled by timeout (60 s)","1773617923000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 50.31s\nKilled by timeout (60 s)","1773615765000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 48.49s\nKilled by timeout (60 s)","1773615308000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 48.26s\nKilled by timeout (60 s)","1773608284000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 40.00s\nKilled by timeout (60 s)","1773606561000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 46.07s\nKilled by timeout (60 s)","1773592886000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 45.80s\nKilled by timeout (60 s)","1773572070000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 46.89s\nKilled by timeout (60 s)","1773570977000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 45.80s\nKilled by timeout (60 s)","1773562367000000":"1 test: 1 - FAIL\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (fail) duration: 51.89s\nKilled by timeout (60 s)","1773562210000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 48.86s\nKilled by timeout (60 s)","1773547892000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 48.86s\nKilled by timeout (60 s)","1773526956000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 49.22s\nKilled by timeout (60 s)","1773519197000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 49.32s\nKilled by timeout (60 s)","1773508856000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.79s\nKilled by timeout (60 s)","1773495254000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.27s\nKilled by timeout (60 s)","1773495175000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 51.28s\nKilled by timeout (60 s)","1773489674000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 48.73s\nKilled by timeout (60 s)","1773488352000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 54.10s\nKilled by timeout (60 s)","1773488108000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 45.87s\nKilled by timeout (60 s)","1773485425000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.65s\nKilled by timeout (60 s)","1773485307000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 72.69s","1773485102000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 51.35s\nKilled by timeout (60 s)","1773481064000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 86.42s","1773428652000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 48.34s\nKilled by timeout (60 s)","1773427126000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 55.24s\nKilled by timeout (60 s)","1773423092000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 86.42s","1773422300000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 46.12s\nKilled by timeout (60 s)","1773421818000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.40s\nKilled by timeout (60 s)","1773421534000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 46.62s\nKilled by timeout (60 s)","1773419832000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 41.02s\nKilled by timeout (60 s)","1773419396000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 48.83s\nKilled by timeout (60 s)","1773418466000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 41.32s\nKilled by timeout (60 s)","1773417296000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 79.27s","1773417033000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.09s\nKilled by timeout (60 s)","1773416973000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 69.57s","1773416600000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 39.40s\nKilled by timeout (60 s)","1773416491000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 45.96s\nKilled by timeout (60 s)","1773414132000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 45.25s\nKilled by timeout (60 s)","1773413019000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 42.91s\nKilled by timeout (60 s)","1773412901000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 60s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (timeout) duration: 59.75s","1773412278000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 44.53s\nKilled by timeout (60 s)","1773409896000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 45.22s\nKilled by timeout (60 s)","1773402640000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.70s\nKilled by timeout (60 s)","1773398367000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 50.06s\nKilled by timeout (60 s)","1773397083000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 43.30s\nKilled by timeout (60 s)","1773392201000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 46.03s\nKilled by timeout (60 s)","1773389248000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 47.53s\nKilled by timeout (60 s)","1773366191000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 43.04s\nKilled by timeout (60 s)","1773365943000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 55.54s\nKilled by timeout (60 s)","1773364460000000":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 43.93s\nKilled by timeout (60 s)"},"ydb/tests/olap/scenario/test_alter_tiering.py.TestAlterTiering.test[many_tables]":{"1773717082000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037952, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037952:1:3:255:1:3896:0] Offset: 0 Size: 3896 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037952:1:3:255:1:3896:0] Offset: 0 Size: 3896 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037952:1:3:255:1:3896:0] Offset: 0 Size: 3896 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037952:1:3:255:1:3896:0] Offset: 0 Size: 3896 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037952:1:3:255:1:3896:0] Offset: 0 Size: 3896 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037952:1:3:255:1:3896:0] Offset: 0 Size: 3896 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773698398000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037908, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037908:1:2:255:1:6536:0] Offset: 496 Size: 256 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773687172000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037908, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3784:0] Offset: 0 Size: 3784 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3784:0] Offset: 0 Size: 3784 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3784:0] Offset: 0 Size: 3784 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3784:0] Offset: 0 Size: 3784 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3784:0] Offset: 0 Size: 3784 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3784:0] Offset: 0 Size: 3784 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773687158000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037904, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037904:1:2:255:1:3752:0] Offset: 320 Size: 208 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773685622000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037904, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037904:1:1:255:1:3848:0] Offset: 344 Size: 232 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773685521000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037895, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037895:1:1:255:1:3840:0] Offset: 344 Size: 216 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773683098000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037901, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:2:255:1:5200:0] Offset: 0 Size: 5200 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:2:255:1:5200:0] Offset: 0 Size: 5200 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:2:255:1:5200:0] Offset: 0 Size: 5200 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:2:255:1:5200:0] Offset: 0 Size: 5200 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:2:255:1:5200:0] Offset: 0 Size: 5200 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:2:255:1:5200:0] Offset: 0 Size: 5200 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773682040000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037903, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037903:1:1:255:1:3680:0] Offset: 304 Size: 200 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773681818000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037904, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037904:1:2:255:1:6328:0] Offset: 448 Size: 256 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773679449000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037903, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037903:1:2:255:1:3848:0] Offset: 344 Size: 216 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:621: in __run\n    self.client.bind_storage_pools(self.domain_name, pools, token=root_token)\nydb/tests/library/clients/kikimr_client.py:197: in bind_storage_pools\n    self.invoke(\nydb/tests/library/clients/kikimr_client.py:122: in invoke\n    return callee(request, timeout=self.__timeout)\ncontrib/python/grpcio/py3/grpc/_channel.py:1181: in __call__\n    return _end_unary_response_blocking(state, call, False, None)\ncontrib/python/grpcio/py3/grpc/_channel.py:1006: in _end_unary_response_blocking\n    raise _InactiveRpcError(state)  # pytype: disable=not-instantiable\nE   grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:\nE   \tstatus = StatusCode.UNAVAILABLE\nE   \tdetails = \"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:29527: Failed to connect to remote host: Connection refused\"\nE   \tdebug_error_string = \"UNKNOWN:Error received from peer  {grpc_message:\"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:29527: Failed to connect to remote host: Connection refused\", grpc_status:14, created_time:\"2026-03-16T15:57:50.005267458+00:00\"}\"\nE   >\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/olap/scenario/conftest.py:76: in setup_class\n    cls._ydb_instance = YdbClusterInstance(ydb_endpoint, ydb_database, cls._get_cluster_config())\nydb/tests/olap/scenario/conftest.py:40: in __init__\n    cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/olap/scenario/test-results/py3test/testing_out_stuff/test_alter_tiering.py.TestAlterTiering.test.many_tables/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/olap/scenario/test-results/py3test/testing_out_stuff/test_alter_tiering.py.TestAlterTiering.test.many_tables/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==887282==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7b6227ca2f40 sp 0x7b6227ca2e70 T52)\nE   ==887282==The signal is caused by a WRITE memory access.\nE   ==887282==Hint: address points to the zero page.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   AddressSanitizer:DEADLYSIGNAL\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7f6270bdfac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7f6270c718bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==887282==Register values:\nE   rax = 0x00007cd26faf4b10  rbx = 0x00007cd26faf4b20  rcx = 0x0000000000000000  rdx = 0x00000f9a4df5e962\nE   rdi = 0x00007d426fb406a8  rsi = 0x00000fa84df680d5  rbp = 0x00007b6227ca2f40  rsp = 0x00007b6227ca2e70\nE   r8 = 0x00000f9a4df5e963   r9 = 0x0000000000000000  r10 = 0x00000f6c44b7fb85  r11 = 0x0000000000000000\nE   r12 = 0x00000f9a4df5e964  r13 = 0x00007d426fb40190  r14 = 0x00007d426fb40080  r15 = 0x00007cd26faf4b18\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\nE   Thread T52 (ydbd.System) created by T0 here:\nE   #0 0x0000211121e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan...","1773677952000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037894, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:2:255:1:3944:0] Offset: 0 Size: 3944 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:2:255:1:3944:0] Offset: 0 Size: 3944 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:2:255:1:3944:0] Offset: 0 Size: 3944 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:2:255:1:3944:0] Offset: 0 Size: 3944 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:2:255:1:3944:0] Offset: 0 Size: 3944 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:2:255:1:3944:0] Offset: 0 Size: 3944 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773676288000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037916, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037916:1:1:255:1:3896:0] Offset: 368 Size: 208 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773674966000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037892, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037892:1:1:255:1:3856:0] Offset: 344 Size: 216 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773674896000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037892, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037892:1:1:255:1:3848:0] Offset: 352 Size: 208 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773666734000000":"automatically muted based on rules","1773665446000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037893, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037893:1:1:255:1:4160:0] Offset: 432 Size: 216 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773663833000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037908, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037908:1:1:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773660879000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037923, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037923:1:2:255:1:4016:0] Offset: 376 Size: 232 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773659987000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037903, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037903:1:1:255:1:3832:0] Offset: 0 Size: 3832 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037903:1:1:255:1:3832:0] Offset: 0 Size: 3832 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037903:1:1:255:1:3832:0] Offset: 0 Size: 3832 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037903:1:1:255:1:3832:0] Offset: 0 Size: 3832 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037903:1:1:255:1:3832:0] Offset: 0 Size: 3832 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037903:1:1:255:1:3832:0] Offset: 0 Size: 3832 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773636520000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037894, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037894:1:1:255:1:3864:0] Offset: 352 Size: 208 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773615765000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037894, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037894:1:2:255:1:3816:0] Offset: 344 Size: 208 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773562210000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037896, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037896:1:1:255:1:3632:0] Offset: 288 Size: 208 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773547892000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037896, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037896:1:1:255:1:3632:0] Offset: 288 Size: 208 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773519197000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037899, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:2:255:1:3736:0] Offset: 0 Size: 3736 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:2:255:1:3736:0] Offset: 0 Size: 3736 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:2:255:1:3736:0] Offset: 0 Size: 3736 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:2:255:1:3736:0] Offset: 0 Size: 3736 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:2:255:1:3736:0] Offset: 0 Size: 3736 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:2:255:1:3736:0] Offset: 0 Size: 3736 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773495254000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037901, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:4:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:4:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:4:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:4:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:4:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037901:1:4:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773489674000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037896, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037896:1:1:255:1:3736:0] Offset: 320 Size: 216 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773488572000000":"automatically muted based on rules","1773488352000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037900, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037900:1:1:255:1:3864:0] Offset: 360 Size: 200 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773488108000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037899, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:4024:0] Offset: 0 Size: 4024 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:4024:0] Offset: 0 Size: 4024 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:4024:0] Offset: 0 Size: 4024 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:4024:0] Offset: 0 Size: 4024 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:4024:0] Offset: 0 Size: 4024 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:4024:0] Offset: 0 Size: 4024 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773487377000000":"automatically muted based on rules","1773485102000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037899, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:3904:0] Offset: 0 Size: 3904 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:3904:0] Offset: 0 Size: 3904 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:3904:0] Offset: 0 Size: 3904 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:3904:0] Offset: 0 Size: 3904 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:3904:0] Offset: 0 Size: 3904 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037899:1:1:255:1:3904:0] Offset: 0 Size: 3904 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773428448000000":"automatically muted based on rules","1773426018000000":"automatically muted based on rules","1773421818000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037914, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037914:1:1:255:1:3824:0] Offset: 0 Size: 3824 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037914:1:1:255:1:3824:0] Offset: 0 Size: 3824 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037914:1:1:255:1:3824:0] Offset: 0 Size: 3824 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037914:1:1:255:1:3824:0] Offset: 0 Size: 3824 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037914:1:1:255:1:3824:0] Offset: 0 Size: 3824 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037914:1:1:255:1:3824:0] Offset: 0 Size: 3824 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773419396000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037910, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037910:1:3:255:1:3696:0] Offset: 0 Size: 3696 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037910:1:3:255:1:3696:0] Offset: 0 Size: 3696 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037910:1:3:255:1:3696:0] Offset: 0 Size: 3696 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037910:1:3:255:1:3696:0] Offset: 0 Size: 3696 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037910:1:3:255:1:3696:0] Offset: 0 Size: 3696 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037910:1:3:255:1:3696:0] Offset: 0 Size: 3696 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773418484000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037892, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3928:0] Offset: 0 Size: 3928 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3928:0] Offset: 0 Size: 3928 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3928:0] Offset: 0 Size: 3928 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3928:0] Offset: 0 Size: 3928 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3928:0] Offset: 0 Size: 3928 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3928:0] Offset: 0 Size: 3928 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773418209000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037892, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3864:0] Offset: 0 Size: 3864 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3864:0] Offset: 0 Size: 3864 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3864:0] Offset: 0 Size: 3864 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3864:0] Offset: 0 Size: 3864 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3864:0] Offset: 0 Size: 3864 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037892:1:1:255:1:3864:0] Offset: 0 Size: 3864 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773417840000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037897, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037897:1:2:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037897:1:2:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037897:1:2:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037897:1:2:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037897:1:2:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037897:1:2:255:1:3848:0] Offset: 0 Size: 3848 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773417033000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037927, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037927:1:2:255:1:5184:0] Offset: 0 Size: 5184 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037927:1:2:255:1:5184:0] Offset: 0 Size: 5184 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037927:1:2:255:1:5184:0] Offset: 0 Size: 5184 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037927:1:2:255:1:5184:0] Offset: 0 Size: 5184 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037927:1:2:255:1:5184:0] Offset: 0 Size: 5184 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037927:1:2:255:1:5184:0] Offset: 0 Size: 5184 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773407147000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037927, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037927:1:2:255:1:6488:0] Offset: 512 Size: 240 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773402640000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037895, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier0::{ Blob: DS:4294967295:[72075186224037895:1:2:255:1:3808:0] Offset: 0 Size: 3808 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier0::{ Blob: DS:4294967295:[72075186224037895:1:2:255:1:3808:0] Offset: 0 Size: 3808 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier0::{ Blob: DS:4294967295:[72075186224037895:1:2:255:1:3808:0] Offset: 0 Size: 3808 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier0::{ Blob: DS:4294967295:[72075186224037895:1:2:255:1:3808:0] Offset: 0 Size: 3808 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier0::{ Blob: DS:4294967295:[72075186224037895:1:2:255:1:3808:0] Offset: 0 Size: 3808 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier0::{ Blob: DS:4294967295:[72075186224037895:1:2:255:1:3808:0] Offset: 0 Size: 3808 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773397572000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037928, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037928:1:2:255:1:3872:0] Offset: 0 Size: 3872 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037928:1:2:255:1:3872:0] Offset: 0 Size: 3872 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037928:1:2:255:1:3872:0] Offset: 0 Size: 3872 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037928:1:2:255:1:3872:0] Offset: 0 Size: 3872 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037928:1:2:255:1:3872:0] Offset: 0 Size: 3872 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037928:1:2:255:1:3872:0] Offset: 0 Size: 3872 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773391749000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037894, reason: task_error:cannot build duplicate filter: [cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:1:255:1:3712:0] Offset: 0 Size: 3712 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:1:255:1:3712:0] Offset: 0 Size: 3712 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:1:255:1:3712:0] Offset: 0 Size: 3712 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:1:255:1:3712:0] Offset: 0 Size: 3712 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:1:255:1:3712:0] Offset: 0 Size: 3712 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;cannot read blob range: /Root/olap_yatests/TestAlterTiering/many_tables/tier1::{ Blob: DS:4294967295:[72075186224037894:1:1:255:1:3712:0] Offset: 0 Size: 3712 }::cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached;]\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773391647000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037901, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037901:1:3:255:1:6376:0] Offset: 464 Size: 232 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')","1773376892000000":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037905, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037905:1:1:255:1:3808:0] Offset: 336 Size: 216 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')"},"ydb/tests/functional/statistics/test_analyze.py.test_basic":{"1773722559000000":"Killed by timeout (600 s)","1773720290000000":"Killed by timeout (600 s)","1773719674000000":"Killed by timeout (600 s)","1773717082000000":"Killed by timeout (600 s)","1773708708000000":"Killed by timeout (600 s)","1773702149000000":"Killed by timeout (600 s)","1773694511000000":"Killed by timeout (600 s)","1773692039000000":"Killed by timeout (600 s)","1773690853000000":"Killed by timeout (600 s)","1773690098000000":"Killed by timeout (600 s)","1773689559000000":"Killed by timeout (600 s)","1773689025000000":"Killed by timeout (600 s)","1773688476000000":"Killed by timeout (600 s)","1773687172000000":"Killed by timeout (600 s)","1773687167000000":"Killed by timeout (600 s)","1773687158000000":"Killed by timeout (600 s)","1773687146000000":"Killed by timeout (600 s)","1773685622000000":"Killed by timeout (600 s)","1773683098000000":"Killed by timeout (600 s)","1773681818000000":"Killed by timeout (600 s)","1773680250000000":"Killed by timeout (600 s)","1773679933000000":"Killed by timeout (600 s)","1773679449000000":"Killed by timeout (600 s)","1773679007000000":"Killed by timeout (600 s)","1773678560000000":"Killed by timeout (600 s)","1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:621: in __run\n    self.client.bind_storage_pools(self.domain_name, pools, token=root_token)\nydb/tests/library/clients/kikimr_client.py:197: in bind_storage_pools\n    self.invoke(\nydb/tests/library/clients/kikimr_client.py:122: in invoke\n    return callee(request, timeout=self.__timeout)\ncontrib/python/grpcio/py3/grpc/_channel.py:1181: in __call__\n    return _end_unary_response_blocking(state, call, False, None)\ncontrib/python/grpcio/py3/grpc/_channel.py:1006: in _end_unary_response_blocking\n    raise _InactiveRpcError(state)  # pytype: disable=not-instantiable\nE   grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:\nE   \tstatus = StatusCode.UNAVAILABLE\nE   \tdetails = \"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:21546: Failed to connect to remote host: Connection refused\"\nE   \tdebug_error_string = \"UNKNOWN:Error received from peer  {created_time:\"2026-03-16T15:31:37.529176494+00:00\", grpc_status:14, grpc_message:\"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:21546: Failed to connect to remote host: Connection refused\"}\"\nE   >\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/library/fixtures/__init__.py:54: in ydb_cluster\n    cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/statistics/test-results/py3test/testing_out_stuff/test_analyze.py.test_basic/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/statistics/test-results/py3test/testing_out_stuff/test_analyze.py.test_basic/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==509727==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7b4ea50e4f40 sp 0x7b4ea50e4e70 T52)\nE   ==509727==The signal is caused by a WRITE memory access.\nE   ==509727==Hint: address points to the zero page.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   AddressSanitizer:DEADLYSIGNAL\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7f4eee096ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7f4eee1288bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==509727==Register values:\nE   rax = 0x00007cbeecf45690  rbx = 0x00007cbeecf456a0  rcx = 0x0000000000000000  rdx = 0x00000f97dd9e8ad2\nE   rdi = 0x00007d2eecf5fea8  rsi = 0x00000fa5dd9ebfd5  rbp = 0x00007b4ea50e4f40  rsp = 0x00007b4ea50e4e70\nE   r8 = 0x00000f97dd9e8ad3   r9 = 0x0000000000000000  r10 = 0x00000f69d4603885  r11 = 0x0000000000000000\nE   r12 = 0x00000f97dd9e8ad4  r13 = 0x00007d2eecf5f990  r14 = 0x00007d2eecf5f880  r15 = 0x00007cbeecf45698\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\nE   Thread T52 (ydbd.System) created by T0 here:\nE   #0 0x0000211121e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149e555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149e555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf50ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 0x00...","1773678224000000":"Killed by timeout (600 s)","1773677952000000":"Killed by timeout (600 s)","1773677718000000":"Killed by timeout (600 s)","1773677510000000":"Killed by timeout (600 s)","1773677230000000":"Killed by timeout (600 s)","1773677073000000":"Killed by timeout (600 s)","1773676288000000":"Killed by timeout (600 s)","1773675427000000":"Killed by timeout (600 s)","1773675081000000":"Killed by timeout (600 s)","1773674966000000":"Killed by timeout (600 s)","1773674896000000":"Killed by timeout (600 s)","1773674709000000":"Killed by timeout (600 s)","1773674432000000":"Killed by timeout (600 s)","1773672978000000":"Killed by timeout (600 s)","1773672747000000":"Killed by timeout (600 s)","1773668400000000":"Killed by timeout (600 s)","1773667338000000":"Killed by timeout (600 s)","1773667038000000":"Killed by timeout (600 s)","1773665446000000":"Killed by timeout (600 s)","1773663833000000":"Killed by timeout (600 s)","1773660879000000":"Killed by timeout (600 s)","1773659987000000":"Killed by timeout (600 s)","1773659879000000":"Killed by timeout (600 s)","1773655409000000":"Killed by timeout (600 s)","1773653768000000":"Killed by timeout (600 s)","1773651160000000":"Killed by timeout (600 s)","1773644268000000":"Killed by timeout (600 s)","1773643731000000":"Killed by timeout (600 s)","1773642105000000":"Killed by timeout (600 s)","1773636520000000":"Killed by timeout (600 s)","1773634122000000":"Killed by timeout (600 s)","1773627098000000":"Killed by timeout (600 s)","1773626829000000":"Killed by timeout (600 s)","1773622583000000":"Killed by timeout (600 s)","1773617923000000":"Killed by timeout (600 s)","1773615308000000":"Killed by timeout (600 s)","1773608284000000":"Killed by timeout (600 s)","1773592886000000":"Killed by timeout (600 s)","1773572070000000":"Killed by timeout (600 s)","1773570977000000":"Killed by timeout (600 s)","1773562367000000":"Killed by timeout (600 s)","1773562210000000":"Killed by timeout (600 s)","1773547892000000":"Killed by timeout (600 s)","1773519197000000":"Killed by timeout (600 s)","1773517997000000":"Killed by timeout (600 s)","1773510184000000":"Killed by timeout (600 s)","1773508856000000":"Killed by timeout (600 s)","1773495175000000":"Killed by timeout (600 s)","1773494895000000":"Killed by timeout (600 s)","1773489674000000":"Killed by timeout (600 s)","1773488352000000":"Killed by timeout (600 s)","1773488108000000":"Killed by timeout (600 s)","1773485812000000":"Killed by timeout (600 s)","1773485425000000":"Killed by timeout (600 s)","1773485307000000":"Killed by timeout (600 s)","1773481064000000":"Killed by timeout (600 s)","1773463643000000":"Killed by timeout (600 s)","1773428652000000":"Killed by timeout (600 s)","1773427126000000":"Killed by timeout (600 s)","1773423092000000":"Killed by timeout (600 s)","1773422300000000":"Killed by timeout (600 s)","1773421818000000":"Killed by timeout (600 s)","1773421534000000":"Killed by timeout (600 s)","1773419832000000":"Killed by timeout (600 s)","1773419396000000":"Killed by timeout (600 s)","1773419081000000":"Killed by timeout (600 s)","1773418484000000":"Killed by timeout (600 s)","1773418466000000":"Killed by timeout (600 s)","1773418209000000":"Killed by timeout (600 s)","1773417667000000":"Killed by timeout (600 s)","1773417296000000":"Killed by timeout (600 s)","1773417175000000":"ydb/tests/functional/statistics/test_analyze.py:32: in test_basic\n    session_pool.execute_with_retries(f'''\ncontrib/python/ydb/py3/ydb/query/pool.py:259: in execute_with_retries\n    return retry_operation_sync(wrapped_callee, retry_settings)\ncontrib/python/ydb/py3/ydb/retries.py:163: in retry_operation_sync\n    for next_opt in opt_generator:\ncontrib/python/ydb/py3/ydb/retries.py:153: in retry_operation_impl\n    raise status\ncontrib/python/ydb/py3/ydb/retries.py:118: in retry_operation_impl\n    result = YdbRetryOperationFinalResult(callee(*args, **kwargs))\ncontrib/python/ydb/py3/ydb/query/pool.py:255: in wrapped_callee\n    with self.checkout(timeout=retry_settings.max_session_acquire_timeout) as session:\ncontrib/python/ydb/py3/ydb/query/pool.py:340: in __enter__\n    self._session = self._pool.acquire(self._timeout)\ncontrib/python/ydb/py3/ydb/query/pool.py:117: in acquire\n    session = self._create_new_session(time_left)\ncontrib/python/ydb/py3/ydb/query/pool.py:69: in _create_new_session\n    session.create(settings=BaseRequestSettings().with_timeout(timeout))\ncontrib/python/ydb/py3/ydb/query/session.py:371: in create\n    self._create_call(settings=settings)\ncontrib/python/ydb/py3/ydb/query/session.py:160: in _create_call\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:469: in __call__\n    connection = self._store.get(preferred_endpoint)\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:153: in get\n    raise issues.ConnectionLost(\"Couldn't find valid connection\")\nE   ydb.issues.ConnectionLost: Couldn't find valid connectionteardown failed:\nydb/tests/library/fixtures/__init__.py:117: in ydb_database\n    yield from _ydb_database(ydb_cluster, ydb_root, ydb_safe_test_name)\nydb/tests/library/fixtures/__init__.py:111: in _ydb_database\n    with ydb_database_ctx(cluster, database):\ncontrib/tools/python3/Lib/contextlib.py:148: in __exit__\n    next(self.gen)\nydb/tests/library/fixtures/__init__.py:103: in ydb_database_ctx\n    ydb_cluster.unregister_and_stop_slots(database_nodes)\nydb/tests/library/harness/kikimr_runner.py:726: in unregister_and_stop_slots\n    i.stop()\nydb/tests/library/harness/kikimr_runner.py:303: in stop\n    super(KiKiMRNode, self).stop()\nydb/tests/library/harness/daemon.py:194: in stop\n    if not self.__check_can_launch_stop(\"stop\"):\nydb/tests/library/harness/daemon.py:171: in __check_can_launch_stop\n    raise DaemonError(\nE   ydb.tests.library.harness.daemon.DaemonError: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/statistics/test-results/py3test/testing_out_stuff/test_analyze.py.test_basic/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/statistics/test-results/py3test/testing_out_stuff/test_analyze.py.test_basic/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T14:46:29.372227Z): verification=IndexConstructor.DeserializeFromProto(indexSchema);fline=update.cpp:24;incorrect_proto=Name: \"__minmax_key\"\nE   MinMaxIndex {\nE   ColumnName: \"key\"\nE   }\nE   ;\nE   ydb/library/actors/core/log.cpp:857\nE   ~TVerifyFormattedRecordWriter(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214A5CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x2149355A\nE   2. /tmp//-S/ydb/library/actors/core/log.cpp:857: ~TVerifyFormattedRecordWriter @ 0x23D4A6BD\nE   3. /tmp//-S/ydb/core/tx/schemeshard/olap/indexes/update.cpp:24: DeserializeFromProto @ 0x2BA10B2B\nE   4. /tmp//-S/ydb/core/tx/schemeshard/olap/indexes/update.cpp:38: Parse @ 0x2BA11E35\nE   5. /tmp//-S/ydb/core/tx/schemeshard/olap/schema/schema.cpp:106: AddDefaultMinMaxIndexes @ 0x2BA145D8\nE   6. /tmp//-S/ydb/core/tx/schemeshard/olap/operations/create_table.cpp:198: DoDeserialize @ 0x2B9D7206\nE   7. /tmp//-S/ydb/core/tx/schemeshard/olap/operations/create_table.cpp:33: Deserialize @ 0x2B9CC7F1\nE   8. /tmp//-S/ydb/core/tx/schemeshard/olap/operations/create_table.cpp:55: BuildTableInfo @ 0x2B9CC7F1\nE   9. /tmp//-S/ydb/core/tx/schemeshard/olap/operations/create_table.cpp:711: Propose @ 0x2B9C5BBD\nE   10. /tmp//-S/ydb/core/tx/schemeshard/schemeshard__operation.cpp:117: ProcessOperationParts @ 0x2B8A7E66\nE   11. /tmp//-S/ydb/core/tx/schemeshard/schemeshard__operation.cpp:292: IgniteOperation @ 0x2B8AE2F7\nE   12. /tmp//-S/ydb/core/tx/schemeshard/schemeshard__operation.cpp:413: Execute @ 0x2B948E78\nE   13. /tmp//-S/ydb/core/tablet_flat/flat_executor.cpp:2036: ExecuteTransaction @ 0x28F4F3B0\nE   14. /tmp//-S/ydb/core/tablet_flat/flat_executor.cpp:1950: DoExecute @ 0x28F4970A\nE   15. /tmp//-S/ydb/core/tablet_flat/flat_executor.cpp:1964: Execute @ 0x28F51FDE\nE   16. /tmp//-S/ydb/core/tablet_flat/tablet_flat_executed.cpp:62: Execute @ 0x28EE498A\nE   17. /tmp//-S/ydb/core/tablet_flat/tablet_flat_executed.cpp:57: Execute @ 0x28EE498A\nE   18. /tmp//-S/ydb/core/tx/schemeshard/schemeshard_impl.cpp:6160: Handle @ 0x2C92B319\nE   19. /tmp//-S/ydb/core/tx/schemeshard/schemeshard_impl.cpp:5326: StateWork @ 0x2C86BFF8\nE   20. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C312A7\nE   21. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D231B1\nE   22. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D2CDC6\nE   23. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D2C37D\nE   24. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D2E3EE\nE   25. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214AAAB4\nE   26. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21128506\nE   27. ??:0: ?? @ 0x7FBF83C22AC2\nE   28. ??:0: ?? @ 0x7FBF83CB48BF","1773417033000000":"Killed by timeout (600 s)","1773416913000000":"Killed by timeout (600 s)","1773416491000000":"Killed by timeout (600 s)","1773414132000000":"Killed by timeout (600 s)","1773413019000000":"Killed by timeout (600 s)","1773412901000000":"Killed by timeout (600 s)","1773412877000000":"Killed by timeout (600 s)","1773412278000000":"Killed by timeout (600 s)","1773409896000000":"Killed by timeout (600 s)","1773408162000000":"Killed by timeout (600 s)","1773407266000000":"Killed by timeout (600 s)","1773407147000000":"Killed by timeout (600 s)","1773406369000000":"Killed by timeout (600 s)","1773405522000000":"Killed by timeout (600 s)","1773405071000000":"Killed by timeout (600 s)","1773404601000000":"Killed by timeout (600 s)","1773402640000000":"Killed by timeout (600 s)","1773401100000000":"Killed by timeout (600 s)","1773398367000000":"Killed by timeout (600 s)","1773397572000000":"Killed by timeout (600 s)","1773397083000000":"Killed by timeout (600 s)","1773394916000000":"Killed by timeout (600 s)","1773394033000000":"Killed by timeout (600 s)","1773392201000000":"Killed by timeout (600 s)","1773391749000000":"Killed by timeout (600 s)","1773391647000000":"Killed by timeout (600 s)","1773389368000000":"Killed by timeout (600 s)","1773389248000000":"Killed by timeout (600 s)","1773389101000000":"Killed by timeout (600 s)","1773388600000000":"Killed by timeout (600 s)","1773376892000000":"Killed by timeout (600 s)","1773374483000000":"Killed by timeout (600 s)","1773366191000000":"Killed by timeout (600 s)","1773365943000000":"Killed by timeout (600 s)","1773364460000000":"Killed by timeout (600 s)","1773361816000000":"Killed by timeout (600 s)"},"ydb/tests/functional/audit/test_canonical_records.py.test_dstool_evict_vdisk_grpc":{"1773735610000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:28046', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773729526000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:4828', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773723348000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:15804', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773723207000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:22457', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773722559000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:30202', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773720290000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:1539', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773719674000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:7257', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773717082000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:63305', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773708708000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:4692', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773702149000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:23876', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773702129000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:3317', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773698398000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:27285', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773694511000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:8871', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773692039000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:13157', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773690853000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:30194', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773690441000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:7938', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773690098000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:30659', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773689559000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:9465', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773689025000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:13384', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773688476000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:2578', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773687448000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:9413', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773687172000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:30674', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773687167000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:1209', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773687158000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:20950', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773687146000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:14531', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 5\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773685622000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:64414', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773685521000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:26765', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773683098000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:15535', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773682225000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:15358', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773682040000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:26873', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773681818000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:19205', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773680250000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:7334', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773679933000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:8464', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773679449000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:25551', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773679120000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:3273', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773679007000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:24089', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773678560000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:6587', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:602: in __run\n    self.add_storage_pool(\nydb/tests/library/harness/kikimr_runner.py:965: in add_storage_pool\n    self._bs_config_invoke(request)\nydb/tests/library/harness/kikimr_runner.py:918: in _bs_config_invoke\n    self.__call_kikimr_new_cli(\nydb/tests/library/harness/kikimr_runner.py:490: in __call_kikimr_new_cli\n    return yatest.common.execute(full_command, env=env)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/ydbd/ydbd --server grpc://localhost:19004 admin blobstorage config invoke --proto=Command {\nE     DefineStoragePool {\nE       BoxId: 1\nE       StoragePoolId: 1\nE       Name: \"dynamic_storage_pool:1\"\nE       ErasureSpecies: \"none\"\nE       VDiskKind: \"Default\"\nE       Kind: \"hdd\"\nE       NumGroups: 2\nE       PDiskFilter {\nE         Property {\nE           Type: ROT\nE         }\nE         Property {\nE           Kind: 0\nE         }\nE       }\nE     }\nE   }\nE   ' has failed with code 1.\nE   Errors:\nE   MB-0001 failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:19004: Failed to connect to remote host: Connection refused\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/library/fixtures/__init__.py:54: in ydb_cluster\n    cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==41343==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7b914f592f40 sp 0x7b914f592e70 T52)\nE   ==41343==The signal is caused by a WRITE memory access.\nE   ==41343==Hint: address points to the zero page.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7f91984ceac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7f91985608bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==41343==Register values:\nE   rax = 0x00007d0197435010  rbx = 0x00007d0197435020  rcx = 0x0000000000000000  rdx = 0x00000fa032e86a02\nE   rdi = 0x00007d71974706a8  rsi = 0x00000fae32e8e0d5  rbp = 0x00007b914f592f40  rsp = 0x00007b914f592e70\nE   r8 = 0x00000fa032e86a03   r9 = 0x0000000000000000  r10 = 0x00000f7229a98f85  r11 = 0x0000000000000000\nE   r12 = 0x00000fa032e86a04  r13 = 0x00007d7197470190  r14 = 0x00007d7197470080  r15 = 0x00007d0197435018\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecat...","1773678224000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:32528', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773677952000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:14887', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773677718000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:15837', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773677510000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:25134', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773677230000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:10637', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773677073000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:31915', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773676288000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:16340', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773675427000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:26331', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773675081000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:3139', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773674966000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:23313', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773674896000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:6476', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773674709000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:27669', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773674432000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:2749', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773672978000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:8303', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773672747000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:27499', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773671187000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:28893', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773669292000000":"ydb/tests/functional/audit/test_canonical_records.py:350: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:57: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode('utf-8')}\\n\\nstdout:\\n{proc_result.std_out.decode('utf-8')}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:17234', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, add --verbose for more info\nE   \nE   \nE   stdout:\n\nlogsdir:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff\nlog:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc.log","1773668553000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:3558', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773668400000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:24247', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773668028000000":"ydb/tests/functional/audit/test_canonical_records.py:350: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:57: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode('utf-8')}\\n\\nstdout:\\n{proc_result.std_out.decode('utf-8')}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:19932', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, add --verbose for more info\nE   \nE   \nE   stdout:\n\nlogsdir:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff\nlog:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc.log","1773667338000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:8303', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773667038000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:11418', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773666791000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:28139', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773666734000000":"ydb/tests/functional/audit/test_canonical_records.py:350: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:57: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode('utf-8')}\\n\\nstdout:\\n{proc_result.std_out.decode('utf-8')}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:9165', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, add --verbose for more info\nE   \nE   \nE   stdout:\n\nlogsdir:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff\nlog:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc.log","1773665446000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:24482', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773663833000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:18008', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773661068000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:3558', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773660879000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:63301', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773659987000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:21121', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773659879000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:18480', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773655409000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:19319', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773653768000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:7951', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773653670000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:63161', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773651160000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:27634', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773644268000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:12093', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:1:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:2:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773643731000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:22997', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773642105000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:9554', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773637789000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:28409', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 5\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773637210000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:24177', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773636520000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:1698', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773634122000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:29296', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773627098000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:25072', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 5\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773626829000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:11395', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773622583000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:13405', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773617923000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:11916', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773615765000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:12984', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773615308000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:11892', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773608284000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:13114', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773606561000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:19895', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773592886000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:22152', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773572070000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:23604', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773570977000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:22152', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773562367000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:20247', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773562210000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:14319', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773550734000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:3852', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773550610000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:32524', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773549981000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:27873', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773547892000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:14319', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773526956000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:7074', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773519197000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:61636', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773517997000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:13435', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773510184000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:22774', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773508856000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:63006', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773495254000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:15815', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773495175000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:28799', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773494895000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:22024', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773489674000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:30246', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773488572000000":"ydb/tests/functional/audit/test_canonical_records.py:350: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:57: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode('utf-8')}\\n\\nstdout:\\n{proc_result.std_out.decode('utf-8')}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:19692', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, add --verbose for more info\nE   \nE   \nE   stdout:\n\nlogsdir:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff\nlog:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc.log","1773488352000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:13189', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773488108000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:8926', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773487377000000":"ydb/tests/functional/audit/test_canonical_records.py:350: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:57: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode('utf-8')}\\n\\nstdout:\\n{proc_result.std_out.decode('utf-8')}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:3786', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, add --verbose for more info\nE   \nE   \nE   stdout:\n\nlogsdir:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff\nlog:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc.log","1773486157000000":"ydb/tests/functional/audit/test_canonical_records.py:350: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:57: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode('utf-8')}\\n\\nstdout:\\n{proc_result.std_out.decode('utf-8')}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:29154', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, add --verbose for more info\nE   \nE   \nE   stdout:\n\nlogsdir:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff\nlog:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc.log","1773485812000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:6113', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773485425000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:12703', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773485307000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:19000', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773485102000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:17886', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773481064000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:5467', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773464419000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:5727', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773464285000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:20211', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773463643000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:18628', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773428652000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:24881', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773428448000000":"ydb/tests/functional/audit/test_canonical_records.py:350: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:57: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode('utf-8')}\\n\\nstdout:\\n{proc_result.std_out.decode('utf-8')}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:2263', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, add --verbose for more info\nE   \nE   \nE   stdout:\n\nlogsdir:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff\nlog:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc.log","1773427248000000":"ydb/tests/functional/audit/test_canonical_records.py:350: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:57: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode('utf-8')}\\n\\nstdout:\\n{proc_result.std_out.decode('utf-8')}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:22688', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, add --verbose for more info\nE   \nE   \nE   stdout:\n\nlogsdir:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff\nlog:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc.log","1773427126000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:17408', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773426018000000":"ydb/tests/functional/audit/test_canonical_records.py:350: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:57: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode('utf-8')}\\n\\nstdout:\\n{proc_result.std_out.decode('utf-8')}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:24556', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, add --verbose for more info\nE   \nE   \nE   stdout:\n\nlogsdir:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff\nlog:\n/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/audit/test-results/py3test/testing_out_stuff/test_canonical_records.py.test_dstool_evict_vdisk_grpc.log","1773423092000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:5467', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773422300000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:30649', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773421818000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:23553', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773421534000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:19950', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773419832000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:10219', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773419396000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:28971', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773419081000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:18469', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 5\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773418484000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:10834', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773418466000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:5369', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773418209000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:62206', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773417840000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:63213', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773417667000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:21740', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773417296000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:12626', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773417175000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:4324', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773417033000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:29250', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773416973000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:5546', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773416913000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:6747', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773416723000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:23720', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773416600000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:9522', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773416491000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:12735', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773414132000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:16124', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773413019000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:8539', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773412901000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:21128', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:1:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:2:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773412877000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:2550', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773412278000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:1803', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773409896000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:3582', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773408162000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:31205', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773407266000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:6036', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773407147000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:8696', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773406369000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:21024', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773405522000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:25898', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773404601000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:6127', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773402640000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:12493', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773401100000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:15991', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773398367000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:8402', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773397572000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:27917', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773397083000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:27321', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773394916000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:19470', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773394033000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:2195', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773392201000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:63552', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773391749000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:30238', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773391647000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:5141', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773389368000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:12814', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773389248000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:20479', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773389101000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:16338', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773388600000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:26549', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773378178000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:30452', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 5\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773378027000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:19497', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 5\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773377371000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:9804', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 5\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773376892000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:1352', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773374483000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:26549', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773366191000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:23710', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 5\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773365943000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:5335', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 5\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773364460000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:32207', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","1773361816000000":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:13847', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1001\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:"},"ydb/tests/functional/api/test_discovery.py.TestDiscoveryFaultInjectionSlotStop.test_scenario":{"1773735610000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773723207000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773722559000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773717082000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773702129000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773698398000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773692039000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773690441000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773690345000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773687167000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773683098000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773682040000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773681818000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:621: in __run\n    self.client.bind_storage_pools(self.domain_name, pools, token=root_token)\nydb/tests/library/clients/kikimr_client.py:197: in bind_storage_pools\n    self.invoke(\nydb/tests/library/clients/kikimr_client.py:122: in invoke\n    return callee(request, timeout=self.__timeout)\ncontrib/python/grpcio/py3/grpc/_channel.py:1181: in __call__\n    return _end_unary_response_blocking(state, call, False, None)\ncontrib/python/grpcio/py3/grpc/_channel.py:1006: in _end_unary_response_blocking\n    raise _InactiveRpcError(state)  # pytype: disable=not-instantiable\nE   grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:\nE   \tstatus = StatusCode.UNAVAILABLE\nE   \tdetails = \"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:18475: Failed to connect to remote host: Connection refused\"\nE   \tdebug_error_string = \"UNKNOWN:Error received from peer  {grpc_message:\"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:18475: Failed to connect to remote host: Connection refused\", grpc_status:14, created_time:\"2026-03-16T15:48:34.919455847+00:00\"}\"\nE   >\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/functional/api/test_discovery.py:116: in setup_class\n    cls.cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/api/test-results/py3test/testing_out_stuff/test_discovery.py.TestDiscoveryFaultInjectionSlotStop.test_scenario/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/api/test-results/py3test/testing_out_stuff/test_discovery.py.TestDiscoveryFaultInjectionSlotStop.test_scenario/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==597871==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7b145e119f40 sp 0x7b145e119e70 T52)\nE   ==597871==The signal is caused by a WRITE memory access.\nE   ==597871==Hint: address points to the zero page.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   AddressSanitizer:DEADLYSIGNAL\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7f14a7057ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7f14a70e98bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==597871==Register values:\nE   rax = 0x00007c84a5ef0f90  rbx = 0x00007c84a5ef0fa0  rcx = 0x0000000000000000  rdx = 0x00000f9094bde1f2\nE   rdi = 0x00007cf4a5f112a8  rsi = 0x00000f9e94be2255  rbp = 0x00007b145e119f40  rsp = 0x00007b145e119e70\nE   r8 = 0x00000f9094bde1f3   r9 = 0x0000000000000000  r10 = 0x00000f628b96d485  r11 = 0x0000000000000000\nE   r12 = 0x00000f9094bde1f4  r13 = 0x00007cf4a5f10d90  r14 = 0x00007cf4a5f10c80  r15 = 0x00007c84a5ef0f98\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\nE   Thread T52 (ydbd.System) created by T0 here:\nE   #0 0x0000211121e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149e555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149e555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf50ec in NActors::TBasicExecutorPool::Start() /-S...","1773677718000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773677510000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773677230000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773677073000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773674896000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773674709000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773674432000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773661040000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773653768000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773653670000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773643731000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773637789000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773637210000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773636520000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773634122000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773627098000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773622583000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773608284000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773562367000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773526956000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773519197000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773517997000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773488352000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773488108000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773485812000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773485102000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773419396000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773418466000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773417840000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773417667000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773417175000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773417033000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773416973000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773416913000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773416491000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773413019000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773412901000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773412278000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773409896000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773405522000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773405071000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773397572000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773397083000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773391749000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773389101000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773388600000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773374483000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","1773361816000000":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)"},"ydb/tests/functional/nbs/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed":{"1773722559000000":"teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   ======= terminate() call stack ========\nE   0. /tmp//-S/ydb/core/driver_lib/run/main.cpp:182: KikimrTerminateHandler @ 0x40AF2C89\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: terminate @ 0x211C4B48\nE   2. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:821: report_failure @ 0x211C5F0E\nE   3. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: __cxa_rethrow @ 0x211C5F0E\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C36517\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D281B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D31DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3137D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D333EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214AFB54\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2112D5A6\nE   11. ??:0: ?? @ 0x7FC81986FAC2\nE   12. ??:0: ?? @ 0x7FC8199018BF\nE   ======== exception call stack =========\nE   0. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:839: throw_exception @ 0x211C536C\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:882: __cxa_throw @ 0x211C518C\nE   2. /-S/contrib/libs/cxxsupp/libcxx/include/stdexcept:251: ?? @ 0x1DB2FD5C\nE   3. /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1787: at @ 0x507E2844\nE   4. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40: ChildSpanEndOk @ 0x507E2844\nE   5. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164: OnWriteFinished @ 0x507E4830\nE   6. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365: operator() @ 0x507B2B55\nE   7. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x507B2B55\nE   8. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x507B2B55\nE   9. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x507B2B55\nE   10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x507B2B55\nE   11. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x507B2B55\nE   12. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x507D6EAB\nE   13. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x507D6EAB\nE   14. /-S/library/cpp/threading/future/core/future-inl.h:210: RunCallbacks @ 0x507D6EAB\nE   15. /-S/library/cpp/threading/future/core/future-inl.h:164: TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x507D6C0C\nE   16. /-S/library/cpp/threading/future/core/future-inl.h:136: SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x507C2F73\nE   17. /-S/library/cpp/threading/future/core/future-inl.h:815: SetValue @ 0x507C2F73\nE   18. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329: HandleWritePersistentBufferResult @ 0x507C2F73\nE   19. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653: StateWork @ 0x507BFE2F\nE   20. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C362A7\nE   21. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:1008: __cxa_rethrow @ 0x211C5DC6\nE   22. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C36517\nE   23. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D281B1\nE   24. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D31DC6\nE   25. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3137D\nE   26. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D333EE\nE   27. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214AFB54\nE   28. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2112D5A6\nE   29. ??:0: ?? @ 0x7FC81986FAC2\nE   30. ??:0: ?? @ 0x7FC8199018BF\nE   =======================================\nE   Terminating due to uncaught exception 0x7d68188c4090    what() -> \"unordered_map::at: key not found\"\nE   of type std::out_of_range","1773717082000000":"Killed by timeout (600 s)","1773702129000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773689559000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773687448000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773682040000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksRead': 1370, 'BlocksWritten': 1463, 'EndTime': 1773676864698511, 'Iops': 257, 'ReadLatency': {'Max': 425727, 'Mean': 34223.5562, 'Min': 713, 'P50': 26799, 'P90': 65439, 'P95': 82687, 'P99': 138495, 'P999': 408063, 'StdDeviation': 36960.0602}, 'RequestsCompleted': 2833, 'StartTime': 1773676853697218, 'WriteLatency': {'Max': 541695, 'Mean': 64526.22146, 'Min': 2556, 'P50': 53823, 'P90': 114943, 'P95': 143103, 'P99': 390911, 'P999': 509439, 'StdDeviation': 59622.50163}}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==130207==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x000050831b97 bp 0x7b583fb3ef50 sp 0x7b583fb3ef30 T98)\nE   ==130207==The signal is caused by a READ memory access.\nE   ==130207==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   Current KQP shutdown state: spent 0 seconds, 0 sessions to shutdown\nE   #0 0x000050831b97 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x000050831b97 in find /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1393:78\nE   #2 0x000050831b97 in at /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1785:18\nE   #3 0x000050831b97 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40:39\nE   #4 0x000050833c30 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x000050801f55 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x000050801f55 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x000050801f55 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x000050801f55 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x000050801f55 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x000050801f55 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000508262ab in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000508262ab in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000508262ab in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x00005082600c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x000050812373 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x000050812373 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x000050812373 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x00005080f22f in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c612a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d531b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d5cdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d5c37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d5e3ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214dab54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x0000211585a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7f5887ddeac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7f5887e708bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==130207==Register values:\nE   rax = 0xbebebebebebebebe  rbx = 0x00000f8b10da0e56  rcx = 0x0000000000000000  rdx = 0x17d7d7d7d7d7d7d7\nE   rdi = 0x00007c5886d07278  rsi = 0x00000000000000f4  rbp = 0x00007b583fb3ef50  rsp = 0x00007b583fb3ef30\nE   r8 = 0x0000000000000002   r9 = 0x0000000000000001  r10 = 0x00000f6b07f67df3  r11 = 0x0000000000000000\nE   r12 = 0x00000f6b87f5fdf4  r13 = 0x00007c5886d072b0  r14 = 0x00007c5886d07258  r15 = 0x00000000000000f4\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table in find<unsigned long>\nE   Thread T98 (ydbd.System) created by T0 here:\nE   #0 0x00002113f1e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x0000214cb555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x0000214cb555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023d220ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 0x000023cbc223 in NActors::T...","1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:621: in __run\n    self.client.bind_storage_pools(self.domain_name, pools, token=root_token)\nydb/tests/library/clients/kikimr_client.py:197: in bind_storage_pools\n    self.invoke(\nydb/tests/library/clients/kikimr_client.py:122: in invoke\n    return callee(request, timeout=self.__timeout)\ncontrib/python/grpcio/py3/grpc/_channel.py:1181: in __call__\n    return _end_unary_response_blocking(state, call, False, None)\ncontrib/python/grpcio/py3/grpc/_channel.py:1006: in _end_unary_response_blocking\n    raise _InactiveRpcError(state)  # pytype: disable=not-instantiable\nE   grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:\nE   \tstatus = StatusCode.UNAVAILABLE\nE   \tdetails = \"failed to connect to all addresses; last error: UNKNOWN: ipv6:%5B::1%5D:25634: Failed to connect to remote host: Connection refused\"\nE   \tdebug_error_string = \"UNKNOWN:Error received from peer  {grpc_message:\"failed to connect to all addresses; last error: UNKNOWN: ipv6:%5B::1%5D:25634: Failed to connect to remote host: Connection refused\", grpc_status:14, created_time:\"2026-03-16T15:09:23.441839969+00:00\"}\"\nE   >\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/functional/nbs/common.py:45: in setup\n    self.cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==230800==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7b5cdb391f40 sp 0x7b5cdb391e70 T51)\nE   ==230800==The signal is caused by a WRITE memory access.\nE   ==230800==Hint: address points to the zero page.\nE   AddressSanitizer:DEADLYSIGNAL\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7f5d23ae3ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7f5d23b758bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==230800==Register values:\nE   rax = 0x00007ccd22a41790  rbx = 0x00007ccd22a417a0  rcx = 0x0000000000000000  rdx = 0x00000f99a45482f2\nE   rdi = 0x00007d3d22a21aa8  rsi = 0x00000fa7a4544355  rbp = 0x00007b5cdb391f40  rsp = 0x00007b5cdb391e70\nE   r8 = 0x00000f99a45482f3   r9 = 0x0000000000000000  r10 = 0x00000f6b9b05b885  r11 = 0x0000000000000000\nE   r12 = 0x00000f99a45482f4  r13 = 0x00007d3d22a21590  r14 = 0x00007d3d22a21480  r15 = 0x00007ccd22a41798\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\nE   Thread T51 (ydbd.System) created by T0 here:\nE   #0 0x0000211121e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149e555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149e555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf50ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.c...","1773677952000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773677510000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773674432000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773653670000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773550734000000":"teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   ======= terminate() call stack ========\nE   0. /tmp//-S/ydb/core/driver_lib/run/main.cpp:182: KikimrTerminateHandler @ 0x40AD8BE9\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: terminate @ 0x211C2AA8\nE   2. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:821: report_failure @ 0x211C3E6E\nE   3. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: __cxa_rethrow @ 0x211C3E6E\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C34517\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D261B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D2FDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D2F37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D313EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214ADAB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2112B506\nE   11. ??:0: ?? @ 0x7F8A4CD93AC2\nE   12. ??:0: ?? @ 0x7F8A4CE258BF\nE   ======== exception call stack =========\nE   0. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:839: throw_exception @ 0x211C32CC\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:882: __cxa_throw @ 0x211C30EC\nE   2. /-S/contrib/libs/cxxsupp/libcxx/include/stdexcept:251: ?? @ 0x1DB2DD5C\nE   3. /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1787: at @ 0x507BDFF4\nE   4. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40: ChildSpanEndOk @ 0x507BDFF4\nE   5. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164: OnWriteFinished @ 0x507BFFE0\nE   6. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365: operator() @ 0x5078E305\nE   7. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x5078E305\nE   8. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x5078E305\nE   9. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x5078E305\nE   10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x5078E305\nE   11. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x5078E305\nE   12. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x507B265B\nE   13. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x507B265B\nE   14. /-S/library/cpp/threading/future/core/future-inl.h:210: RunCallbacks @ 0x507B265B\nE   15. /-S/library/cpp/threading/future/core/future-inl.h:164: TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x507B23BC\nE   16. /-S/library/cpp/threading/future/core/future-inl.h:136: SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x5079E723\nE   17. /-S/library/cpp/threading/future/core/future-inl.h:815: SetValue @ 0x5079E723\nE   18. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329: HandleWritePersistentBufferResult @ 0x5079E723\nE   19. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653: StateWork @ 0x5079B5DF\nE   20. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C342A7\nE   21. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:1008: __cxa_rethrow @ 0x211C3D26\nE   22. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C34517\nE   23. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D261B1\nE   24. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D2FDC6\nE   25. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D2F37D\nE   26. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D313EE\nE   27. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214ADAB4\nE   28. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2112B506\nE   29. ??:0: ?? @ 0x7F8A4CD93AC2\nE   30. ??:0: ?? @ 0x7F8A4CE258BF\nE   =======================================\nE   Terminating due to uncaught exception 0x7d2a4bdeaa90    what() -> \"unordered_map::at: key not found\"\nE   of type std::out_of_range","1773550610000000":"teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==1654652==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x0000507bdf47 bp 0x7b27f6ca7f50 sp 0x7b27f6ca7f30 T97)\nE   ==1654652==The signal is caused by a READ memory access.\nE   ==1654652==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   Current KQP shutdown state: spent 0 seconds, not started yet\nE   #0 0x0000507bdf47 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x0000507bdf47 in find /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1393:78\nE   #2 0x0000507bdf47 in at /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1785:18\nE   #3 0x0000507bdf47 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40:39\nE   #4 0x0000507bffe0 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x00005078e305 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x00005078e305 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x00005078e305 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x00005078e305 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x00005078e305 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x00005078e305 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507b265b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507b265b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507b265b in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507b23bc in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x00005079e723 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x00005079e723 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x00005079e723 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x00005079b5df in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214adab4 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x00002112b506 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7f283fd2cac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7f283fdbe8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==1654652==Register values:\nE   rax = 0xbebebebebebebebe  rbx = 0x00000f8507da1e56  rcx = 0x0000000000000001  rdx = 0x17d7d7d7d7d7d7d7\nE   rdi = 0x00007c283ed0f278  rsi = 0x0000000000000501  rbp = 0x00007b27f6ca7f50  rsp = 0x00007b27f6ca7f30\nE   r8 = 0x0000000000000002   r9 = 0x0000000000000001  r10 = 0x00000f64fed94ff3  r11 = 0x0000000000000000\nE   r12 = 0x00000f657ed8cff4  r13 = 0x00007c283ed0f2b0  r14 = 0x00007c283ed0f258  r15 = 0x0000000000000501\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table in find<unsigned long>\nE   Thread T97 (ydbd.System) created by T0 here:\nE   #0 0x000021112141 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149e4b5 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149e4b5 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf50ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 0x000023c8f223 in NActors::TCpu...","1773549981000000":"teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   ======= terminate() call stack ========\nE   Current KQP shutdown state: spent 0 seconds, not started yet\nE   0. /tmp//-S/ydb/core/driver_lib/run/main.cpp:182: KikimrTerminateHandler @ 0x40AD8BE9\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: terminate @ 0x211C2AA8\nE   2. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:821: report_failure @ 0x211C3E6E\nE   3. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: __cxa_rethrow @ 0x211C3E6E\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C34517\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D261B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D2FDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D2F37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D313EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214ADAB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2112B506\nE   11. ??:0: ?? @ 0x7FD470C30AC2\nE   12. ??:0: ?? @ 0x7FD470CC28BF\nE   ======== exception call stack =========\nE   0. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:839: throw_exception @ 0x211C32CC\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:882: __cxa_throw @ 0x211C30EC\nE   2. /-S/contrib/libs/cxxsupp/libcxx/include/stdexcept:251: ?? @ 0x1DB2DD5C\nE   3. /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1787: at @ 0x507BDFF4\nE   4. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40: ChildSpanEndOk @ 0x507BDFF4\nE   5. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164: OnWriteFinished @ 0x507BFFE0\nE   6. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365: operator() @ 0x5078E305\nE   7. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x5078E305\nE   8. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x5078E305\nE   9. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x5078E305\nE   10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x5078E305\nE   11. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x5078E305\nE   12. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x507B265B\nE   13. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x507B265B\nE   14. /-S/library/cpp/threading/future/core/future-inl.h:210: RunCallbacks @ 0x507B265B\nE   15. /-S/library/cpp/threading/future/core/future-inl.h:164: TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x507B23BC\nE   16. /-S/library/cpp/threading/future/core/future-inl.h:136: SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x5079E723\nE   17. /-S/library/cpp/threading/future/core/future-inl.h:815: SetValue @ 0x5079E723\nE   18. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329: HandleWritePersistentBufferResult @ 0x5079E723\nE   19. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653: StateWork @ 0x5079B5DF\nE   20. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C342A7\nE   21. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:1008: __cxa_rethrow @ 0x211C3D26\nE   22. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C34517\nE   23. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D261B1\nE   24. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D2FDC6\nE   25. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D2F37D\nE   26. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D313EE\nE   27. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214ADAB4\nE   28. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2112B506\nE   29. ??:0: ?? @ 0x7FD470C30AC2\nE   30. ??:0: ?? @ 0x7FD470CC28BF\nE   =======================================\nE   Terminating due to uncaught exception 0x7d746fbcec90    what() -> \"unordered_map::at: key not found\"\nE   of type std::out_of_range","1773526956000000":"teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==257151==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x0000507dc197 bp 0x7b607bcbff50 sp 0x7b607bcbff30 T97)\nE   ==257151==The signal is caused by a READ memory access.\nE   ==257151==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   Current KQP shutdown state: spent 0 seconds, not started yet\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x0000507dc197 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x0000507dc197 in find /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1393:78\nE   #2 0x0000507dc197 in at /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1785:18\nE   #3 0x0000507dc197 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40:39\nE   #4 0x0000507de230 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x0000507ac555 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x0000507ac555 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x0000507ac555 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x0000507ac555 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x0000507ac555 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x0000507ac555 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507d08ab in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507d08ab in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507d08ab in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507d060c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x0000507bc973 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x0000507bc973 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x0000507bc973 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x0000507b982f in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c392a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d2b1b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d34dc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d3437d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d363ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214b2ab4 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x000021130506 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7f60c4d8bac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7f60c4e1d8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==257151==Register values:\nE   rax = 0xffffffff00000000  rbx = 0x00000f8c187e39d6  rcx = 0x0000000000000000  rdx = 0x1fffffffe0000000\nE   rdi = 0x00007c60c3f1ce78  rsi = 0x000000000000012a  rbp = 0x00007b607bcbff50  rsp = 0x00007b607bcbff30\nE   r8 = 0x0000000000000002   r9 = 0x0000000000000001  r10 = 0x00000f6c0f797ff3  r11 = 0x0000000000000000\nE   r12 = 0x00000f6c8f78fff4  r13 = 0x00007c60c3f1ceb0  r14 = 0x00007c60c3f1ce58  r15 = 0x000000000000012a\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table in find<unsigned long>\nE   Thread T97 (ydbd.System) created by T0 here:\nE   #0 0x000021117141 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x0000214a34b5 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x0000214a34b5 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cfa0ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 0x000023c94223 in NActors::TCpuMana...","1773494895000000":"Killed by timeout (600 s)","1773418209000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773417840000000":"teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   ======= terminate() call stack ========\nE   Current KQP shutdown state: spent 0 seconds, not started yet\nE   0. /tmp//-S/ydb/core/driver_lib/run/main.cpp:182: KikimrTerminateHandler @ 0x40AD3E99\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: terminate @ 0x211BEAA8\nE   2. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:821: report_failure @ 0x211BFE6E\nE   3. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: __cxa_rethrow @ 0x211BFE6E\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C30517\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D221B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D2BDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D2B37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D2D3EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214A9AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21127506\nE   11. ??:0: ?? @ 0x7F807DBE5AC2\nE   12. ??:0: ?? @ 0x7F807DC778BF\nE   ======== exception call stack =========\nE   0. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:839: throw_exception @ 0x211BF2CC\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:882: __cxa_throw @ 0x211BF0EC\nE   2. /-S/contrib/libs/cxxsupp/libcxx/include/stdexcept:251: ?? @ 0x1DB29D5C\nE   3. /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1787: at @ 0x507BBED4\nE   4. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40: ChildSpanEndOk @ 0x507BBED4\nE   5. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164: OnWriteFinished @ 0x507BDEC0\nE   6. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365: operator() @ 0x5078C1E5\nE   7. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x5078C1E5\nE   8. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x5078C1E5\nE   9. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x5078C1E5\nE   10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x5078C1E5\nE   11. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x5078C1E5\nE   12. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x507B053B\nE   13. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x507B053B\nE   14. /-S/library/cpp/threading/future/core/future-inl.h:210: RunCallbacks @ 0x507B053B\nE   15. /-S/library/cpp/threading/future/core/future-inl.h:164: TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x507B029C\nE   16. /-S/library/cpp/threading/future/core/future-inl.h:136: SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x5079C603\nE   17. /-S/library/cpp/threading/future/core/future-inl.h:815: SetValue @ 0x5079C603\nE   18. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329: HandleWritePersistentBufferResult @ 0x5079C603\nE   19. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653: StateWork @ 0x507994BF\nE   20. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C302A7\nE   21. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:1008: __cxa_rethrow @ 0x211BFD26\nE   22. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C30517\nE   23. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D221B1\nE   24. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D2BDC6\nE   25. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D2B37D\nE   26. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D2D3EE\nE   27. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214A9AB4\nE   28. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21127506\nE   29. ??:0: ?? @ 0x7F807DBE5AC2\nE   30. ??:0: ?? @ 0x7F807DC778BF\nE   =======================================\nE   Terminating due to uncaught exception 0x7d207cd67e90    what() -> \"unordered_map::at: key not found\"\nE   of type std::out_of_range","1773405522000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:51: in test_nbs_load_actor_mixed\n    self.create_disk(disk_id)\nydb/tests/functional/nbs/common.py:102: in create_disk\n    execute_dstool_grpc(\nydb/tests/functional/nbs/helpers.py:38: in execute_dstool_grpc\n    False\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:15899', 'nbs', 'partition', 'create', '--pool', 'ddp1', '--block-size=4096', '--blocks-count=1048576', '--type=ssd', '--disk-id', 'cBwgA2dq6o']\nE    finished with exit code 1, stderr:\nE   \nE   WARNING: failed to fetch data from host localhost:15899 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:15899 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:15899 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:15899 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:15899 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   Connection Error: Can't connect to specified addresses\nE   \nE   \nE   stdout:teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_2/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_2/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T12:13:24.669865Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214BFCC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214AD55A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421F219E\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x4219F921\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C4B2A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3D1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D46DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4637D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D483EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C4AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21142506\nE   11. ??:0: ?? @ 0x7F99165EEAC2\nE   12. ??:0: ?? @ 0x7F99166808BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T12:13:24.672626Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214BFCC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214AD55A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421F219E\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x4219F921\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C4B2A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3D1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D46DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4637D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D483EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C4AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21142506\nE   11. ??:0: ?? @ 0x7F4B39AA0AC2\nE   12. ??:0: ?? @ 0x7F4B39B328BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_7/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_7/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T12:13:24.646001Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214BFCC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214AD55A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421F219E\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x4219F921\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C4B2A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3D1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D46DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4637D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D483EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C4AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21142506\nE   11. ??:0: ?? @ 0x7F359FCA1AC2\nE   12. ??:0: ?? @ 0x7F359FD338BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_8/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_8/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T12:13:24.667036Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214BFCC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214AD55A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421F219E\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x4219F921\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C4B2A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3D1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D46DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4637D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D483EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C4AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21142506\nE   11. ??:0: ?? @ 0x7FAB19CC9AC2\nE   12. ??:0: ?? @ 0x7FAB19D5B8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_9/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/...","1773397083000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:51: in test_nbs_load_actor_mixed\n    self.create_disk(disk_id)\nydb/tests/functional/nbs/common.py:102: in create_disk\n    execute_dstool_grpc(\nydb/tests/functional/nbs/helpers.py:38: in execute_dstool_grpc\n    False\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:1438', 'nbs', 'partition', 'create', '--pool', 'ddp1', '--block-size=4096', '--blocks-count=1048576', '--type=ssd', '--disk-id', 'yEise1lKAF']\nE    finished with exit code 1, stderr:\nE   \nE   WARNING: failed to fetch data from host localhost:1438 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:1438 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:1438 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:1438 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:1438 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   Connection Error: Can't connect to specified addresses\nE   \nE   \nE   stdout:teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T09:47:57.584780Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x21517277\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x21504B0A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x42248ECE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421F6651\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23CA22A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D941B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D9DDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D9D37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D9F3EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x2151C064\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21199AB6\nE   11. ??:0: ?? @ 0x7F7A7EFE0AC2\nE   12. ??:0: ?? @ 0x7F7A7F0728BF","1773389368000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:51: in test_nbs_load_actor_mixed\n    self.create_disk(disk_id)\nydb/tests/functional/nbs/common.py:102: in create_disk\n    execute_dstool_grpc(\nydb/tests/functional/nbs/helpers.py:38: in execute_dstool_grpc\n    False\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:16511', 'nbs', 'partition', 'create', '--pool', 'ddp1', '--block-size=4096', '--blocks-count=1048576', '--type=ssd', '--disk-id', 'p8d1e1pv51']\nE    finished with exit code 1, stderr:\nE   \nE   WARNING: failed to fetch data from host localhost:16511 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:16511 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:16511 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:16511 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:16511 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   Connection Error: Can't connect to specified addresses\nE   \nE   \nE   stdout:teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T07:45:41.583847Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214B7E07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A569A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421E8BFE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x42196381\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C432A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D351B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D3EDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3E37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D403EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BCBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113B506\nE   11. ??:0: ?? @ 0x7F9DBDFE2AC2\nE   12. ??:0: ?? @ 0x7F9DBE0748BF","1773389101000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773388600000000":"Killed by timeout (600 s)","1773377371000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:52: in test_nbs_load_actor_mixed\n    actor_id = self.get_load_actor_adapter_actor_id(disk_id)\nydb/tests/functional/nbs/common.py:130: in get_load_actor_adapter_actor_id\n    assert status == \"success\"\nE   AssertionErrorteardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_2/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_2/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T04:20:41.676875Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214B7E07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A569A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421E8B8E\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x42196311\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C432A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D351B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D3EDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3E37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D403EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BCBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113B506\nE   11. ??:0: ?? @ 0x7FF502118AC2\nE   12. ??:0: ?? @ 0x7FF5021AA8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_3/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_3/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T04:20:41.670054Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214B7E07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A569A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421E8B8E\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x42196311\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C432A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D351B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D3EDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3E37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D403EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BCBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113B506\nE   11. ??:0: ?? @ 0x7F92C776AAC2\nE   12. ??:0: ?? @ 0x7F92C77FC8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_4/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_4/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T04:20:41.742806Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214B7E07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A569A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421E8B8E\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x42196311\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C432A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D351B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D3EDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3E37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D403EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BCBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113B506\nE   11. ??:0: ?? @ 0x7F425F0DAAC2\nE   12. ??:0: ?? @ 0x7F425F16C8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_5/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_5/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T04:20:41.666855Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214B7E07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A569A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421E8B8E\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x42196311\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C432A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D351B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D3EDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3E37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D403EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BCBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113B506\nE   11. ??:0: ?? @ 0x7FF42C74AAC2\nE   12. ??:0: ?? @ 0x7FF42C7DC8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_6/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/...","1773376892000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:51: in test_nbs_load_actor_mixed\n    self.create_disk(disk_id)\nydb/tests/functional/nbs/common.py:102: in create_disk\n    execute_dstool_grpc(\nydb/tests/functional/nbs/helpers.py:38: in execute_dstool_grpc\n    False\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:5124', 'nbs', 'partition', 'create', '--pool', 'ddp1', '--block-size=4096', '--blocks-count=1048576', '--type=ssd', '--disk-id', 'EdjI3egB0Y']\nE    finished with exit code 1, stderr:\nE   \nE   WARNING: failed to fetch data from host localhost:5124 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:5124 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:5124 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:5124 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:5124 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   Connection Error: Can't connect to specified addresses\nE   \nE   \nE   stdout:teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T04:15:07.970497Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214BAE07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A869A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421F247E\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x4219FC01\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C462A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D381B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D41DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4137D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D433EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BFBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113E506\nE   11. ??:0: ?? @ 0x7F2F82602AC2\nE   12. ??:0: ?? @ 0x7F2F826948BF","1773374483000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:51: in test_nbs_load_actor_mixed\n    self.create_disk(disk_id)\nydb/tests/functional/nbs/common.py:102: in create_disk\n    execute_dstool_grpc(\nydb/tests/functional/nbs/helpers.py:38: in execute_dstool_grpc\n    False\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:15054', 'nbs', 'partition', 'create', '--pool', 'ddp1', '--block-size=4096', '--blocks-count=1048576', '--type=ssd', '--disk-id', '46Ai9S6Wz7']\nE    finished with exit code 1, stderr:\nE   \nE   WARNING: failed to fetch data from host localhost:15054 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:15054 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:15054 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:15054 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:15054 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   Connection Error: Can't connect to specified addresses\nE   \nE   \nE   stdout:teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_2/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_2/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T03:32:56.358251Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214C0CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214AE55A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421F39AE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421A1131\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C4C2A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3E1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D47DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4737D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D493EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C5AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21143506\nE   11. ??:0: ?? @ 0x7FAEC52F8AC2\nE   12. ??:0: ?? @ 0x7FAEC538A8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_9/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_9/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T03:32:56.373766Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214C0CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214AE55A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421F39AE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421A1131\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C4C2A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3E1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D47DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4737D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D493EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C5AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21143506\nE   11. ??:0: ?? @ 0x7F7980860AC2\nE   12. ??:0: ?? @ 0x7F79808F28BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T03:32:56.347307Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214C0CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214AE55A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421F39AE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421A1131\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C4C2A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3E1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D47DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4737D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D493EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C5AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21143506\nE   11. ??:0: ?? @ 0x7F8AD28EBAC2\nE   12. ??:0: ?? @ 0x7F8AD297D8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_8/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_8/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T03:32:56.386239Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214C0CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214AE55A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421F39AE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421A1131\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C4C2A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3E1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D47DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4737D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D493EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C5AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21143506\nE   11. ??:0: ?? @ 0x7FA65FFC6AC2\nE   12. ??:0: ?? @ 0x7FA6600588BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_3/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/...","1773364460000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:51: in test_nbs_load_actor_mixed\n    self.create_disk(disk_id)\nydb/tests/functional/nbs/common.py:102: in create_disk\n    execute_dstool_grpc(\nydb/tests/functional/nbs/helpers.py:38: in execute_dstool_grpc\n    False\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:9056', 'nbs', 'partition', 'create', '--pool', 'ddp1', '--block-size=4096', '--blocks-count=1048576', '--type=ssd', '--disk-id', 'uL7VK9vFPJ']\nE    finished with exit code 1, stderr:\nE   \nE   WARNING: failed to fetch data from host localhost:9056 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:9056 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:9056 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:9056 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:9056 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   Connection Error: Can't connect to specified addresses\nE   \nE   \nE   stdout:teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T00:50:50.228813Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214CB7A7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214B903A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x42224ABE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421D2241\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C562A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D481B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D51DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D5137D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D533EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214D0594\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2114EEA6\nE   11. ??:0: ?? @ 0x7F756B61FAC2\nE   12. ??:0: ?? @ 0x7F756B6B18BF"},"ydb/tests/functional/nbs/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read":{"1773735610000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2913, 'EndTime': 1773730628379963, 'Iops': 264, 'RequestsCompleted': 2913, 'StartTime': 1773730617378095, 'WriteLatency': {'Max': 599551, 'Mean': 119979.2111, 'Min': 7316, 'P50': 104127, 'P90': 206079, 'P95': 249983, 'P99': 362239, 'P999': 508927, 'StdDeviation': 67693.63461}}","1773702149000000":"Killed by timeout (600 s)","1773698398000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksRead': 1088, 'EndTime': 1773694307760619, 'Iops': 98, 'ReadLatency': {'Max': 49343, 'Mean': 6459.751838, 'Min': 548, 'P50': 4399, 'P90': 14503, 'P95': 19087, 'P99': 29007, 'P999': 48543, 'StdDeviation': 6394.489061}, 'RequestsCompleted': 1088, 'StartTime': 1773694296759498}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==1561077==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x0000507e2797 bp 0x7b1826c81f50 sp 0x7b1826c81f30 T98)\nE   ==1561077==The signal is caused by a READ memory access.\nE   ==1561077==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x0000507e2797 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x0000507e2797 in find /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1393:78\nE   #2 0x0000507e2797 in at /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1785:18\nE   #3 0x0000507e2797 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40:39\nE   #4 0x0000507e4830 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x0000507b2b55 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x0000507b2b55 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x0000507b2b55 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x0000507b2b55 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x0000507b2b55 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x0000507b2b55 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507d6eab in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507d6eab in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507d6eab in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507d6c0c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x0000507c2f73 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x0000507c2f73 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x0000507c2f73 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x0000507bfe2f in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c362a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d281b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d31dc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d3137d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d333ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214afb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x00002112d5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7f1870512ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7f18705a48bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==1561077==Register values:\nE   rax = 0xbebebebebebebebe  rbx = 0x00000f830dedc876  rcx = 0x0000000000000001  rdx = 0x17d7d7d7d7d7d7d7\nE   rdi = 0x00007c186f6e4378  rsi = 0x00000000000001a1  rbp = 0x00007b1826c81f50  rsp = 0x00007b1826c81f30\nE   r8 = 0x0000000000000002   r9 = 0x0000000000000001  r10 = 0x00000f6304d903f3  r11 = 0x0000000000000000\nE   r12 = 0x00000f6384d883f4  r13 = 0x00007c186f6e43b0  r14 = 0x00007c186f6e4358  r15 = 0x00000000000001a1\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table in find<unsigned long>\nE   Thread T98 (ydbd.System) created by T0 here:\nE   #0 0x0000211141e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x0000214a0555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x0000214a0555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf70ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 0x000023c91223 in NActors::TCpuManager::Start() /-S/ydb/l...","1773692039000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2817, 'EndTime': 1773686850948046, 'Iops': 256, 'RequestsCompleted': 2817, 'StartTime': 1773686839945927, 'WriteLatency': {'Max': 586751, 'Mean': 97788.52893, 'Min': 2132, 'P50': 87167, 'P90': 178175, 'P95': 217215, 'P99': 323327, 'P999': 568319, 'StdDeviation': 69246.07185}}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==162193==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x0000508042d7 bp 0x7ba252d05f20 sp 0x7ba252d05f00 T98)\nE   ==162193==The signal is caused by a READ memory access.\nE   ==162193==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   Current KQP shutdown state: spent 0 seconds, not started yet\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x0000508042d7 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x0000508042d7 in unsigned long std::__y1::__hash_table<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::__unordered_map_hasher<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::hash<unsigned long>, std::__y1::equal_to<unsigned long>, true>, std::__y1::__unordered_map_equal<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::equal_to<unsigned long>, std::__y1::hash<unsigned long>, true>, std::__y1::allocator<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>>>::__erase_unique<unsigned long>(unsigned long const&) /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table:1877:18\nE   #2 0x0000507fb20c in erase /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1336:80\nE   #3 0x0000507fb20c in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:42:26\nE   #4 0x0000507fd0f0 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x0000507cb415 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x0000507cb415 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x0000507cb415 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x0000507cb415 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x0000507cb415 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x0000507cb415 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507ef76b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507ef76b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507ef76b in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507ef4cc in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x0000507db833 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x0000507db833 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x0000507db833 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x0000507d86ef in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c3b2a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d2d1b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d36dc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d3637d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d383ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214b4b54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x0000211325a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7fa29c5a8ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7fa29c63a8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==162193==Register values:\nE   rax = 0x17d7d7d7d7d7d7d7  rbx = 0x00007bd29eb10f48  rcx = 0x0000000000000160  rdx = 0x00000f7a53d621e8\nE   rdi = 0xbebebebebebebebe  rsi = 0x0000000000000000  rbp = 0x00007ba252d05f20  rsp = 0x00007ba252d05f00\nE   r8 = 0x0000000000000001   r9 = 0x0000000000000001  r10 = 0x00007ca29b6fc678  r11 = 0x0000000000000002\nE   r12 = 0x00000f74ca598bf4  r13 = 0x00007ca29b6fc6b0  r14 = 0x00007ca29b6fc658  r15 = 0x0000000000000160\nE  ...","1773690098000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2883, 'EndTime': 1773684716495432, 'Iops': 262, 'RequestsCompleted': 2883, 'StartTime': 1773684705493917, 'WriteLatency': {'Max': 538111, 'Mean': 121740.6264, 'Min': 6780, 'P50': 106303, 'P90': 209535, 'P95': 247935, 'P99': 347903, 'P999': 442879, 'StdDeviation': 65318.1849}}","1773689559000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3004, 'EndTime': 1773684512368185, 'Iops': 272, 'RequestsCompleted': 3004, 'StartTime': 1773684501364200, 'WriteLatency': {'Max': 462847, 'Mean': 116538.8475, 'Min': 7420, 'P50': 103999, 'P90': 198143, 'P95': 244479, 'P99': 325375, 'P999': 398591, 'StdDeviation': 63770.1111}}","1773689025000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2390, 'EndTime': 1773684643899851, 'Iops': 217, 'RequestsCompleted': 2390, 'StartTime': 1773684632897004, 'WriteLatency': {'Max': 582655, 'Mean': 146525.7607, 'Min': 19472, 'P50': 131583, 'P90': 241407, 'P95': 297471, 'P99': 389887, 'P999': 539647, 'StdDeviation': 74901.01413}}","1773687172000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2728, 'EndTime': 1773682056460937, 'Iops': 247, 'RequestsCompleted': 2728, 'StartTime': 1773682045459440, 'WriteLatency': {'Max': 489727, 'Mean': 128223.2405, 'Min': 16912, 'P50': 114559, 'P90': 215679, 'P95': 253311, 'P99': 348159, 'P999': 446207, 'StdDeviation': 64523.41927}}","1773685622000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2706, 'EndTime': 1773680477416497, 'Iops': 245, 'RequestsCompleted': 2706, 'StartTime': 1773680466416181, 'WriteLatency': {'Max': 392447, 'Mean': 129499.2077, 'Min': 17296, 'P50': 118847, 'P90': 215295, 'P95': 256255, 'P99': 323327, 'P999': 373503, 'StdDeviation': 62898.14316}}","1773681818000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2773, 'EndTime': 1773676612880604, 'Iops': 252, 'RequestsCompleted': 2773, 'StartTime': 1773676601878485, 'WriteLatency': {'Max': 472063, 'Mean': 126329.617, 'Min': 12896, 'P50': 113151, 'P90': 209279, 'P95': 252415, 'P99': 344319, 'P999': 403711, 'StdDeviation': 63788.15526}}","1773680169000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with errorteardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   ======= terminate() call stack ========\nE   0. /tmp//-S/ydb/core/driver_lib/run/main.cpp:182: KikimrTerminateHandler @ 0x40ADA3E9\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: terminate @ 0x211C4B48\nE   2. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:821: report_failure @ 0x211C5F0E\nE   3. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: __cxa_rethrow @ 0x211C5F0E\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C36517\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D281B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D31DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3137D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D333EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214AFB54\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2112D5A6\nE   11. ??:0: ?? @ 0x7F123B7ECAC2\nE   12. ??:0: ?? @ 0x7F123B87E8BF\nE   ======== exception call stack =========\nE   0. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:839: throw_exception @ 0x211C536C\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:882: __cxa_throw @ 0x211C518C\nE   2. /-S/contrib/libs/cxxsupp/libcxx/include/stdexcept:251: ?? @ 0x1DB2FD5C\nE   3. /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1787: at @ 0x507C8564\nE   4. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40: ChildSpanEndOk @ 0x507C8564\nE   5. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164: OnWriteFinished @ 0x507CA550\nE   6. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365: operator() @ 0x50798875\nE   7. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x50798875\nE   8. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x50798875\nE   9. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x50798875\nE   10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x50798875\nE   11. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x50798875\nE   12. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x507BCBCB\nE   13. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x507BCBCB\nE   14. /-S/library/cpp/threading/future/core/future-inl.h:210: RunCallbacks @ 0x507BCBCB\nE   15. /-S/library/cpp/threading/future/core/future-inl.h:164: TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x507BC92C\nE   16. /-S/library/cpp/threading/future/core/future-inl.h:136: SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x507A8C93\nE   17. /-S/library/cpp/threading/future/core/future-inl.h:815: SetValue @ 0x507A8C93\nE   18. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329: HandleWritePersistentBufferResult @ 0x507A8C93\nE   19. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653: StateWork @ 0x507A5B4F\nE   20. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C362A7\nE   21. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:1008: __cxa_rethrow @ 0x211C5DC6\nE   22. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C36517\nE   23. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D281B1\nE   24. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D31DC6\nE   25. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3137D\nE   26. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D333EE\nE   27. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214AFB54\nE   28. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2112D5A6\nE   29. ??:0: ?? @ 0x7F123B7ECAC2\nE   30. ??:0: ?? @ 0x7F123B87E8BF\nE   =======================================\nE   Terminating due to uncaught exception 0x7cb23a85e690    what() -> \"unordered_map::at: key not found\"\nE   of type std::out_of_range","1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:595: in __run\n    self.__wait_for_bs_controller_to_start(timeout_seconds=timeout_seconds, token=token)\nydb/tests/library/harness/kikimr_runner.py:982: in __wait_for_bs_controller_to_start\n    bs_controller_started = wait_for(\nydb/tests/library/common/wait_for.py:19: in wait_for\n    time.sleep(step_seconds)\nlibrary/python/pytest/plugins/ya.py:350: in _graceful_shutdown\n    _graceful_shutdown_on_log(not capman.is_globally_capturing())\nlibrary/python/pytest/plugins/ya.py:322: in _graceful_shutdown_on_log\n    pytest.exit(\"Graceful shutdown requested\")\nE   _pytest.outcomes.Exit: Graceful shutdown requested\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/functional/nbs/common.py:45: in setup\n    self.cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==259732==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7bd2c3466f40 sp 0x7bd2c3466e70 T52)\nE   ==259732==The signal is caused by a WRITE memory access.\nE   ==259732==Hint: address points to the zero page.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7fd30c463ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7fd30c4f58bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==259732==Register values:\nE   rax = 0x00007d430b357590  rbx = 0x00007d430b3575a0  rcx = 0x0000000000000000  rdx = 0x00000fa86166aeb2\nE   rdi = 0x00007db30b321aa8  rsi = 0x00000fb661664355  rbp = 0x00007bd2c3466f40  rsp = 0x00007bd2c3466e70\nE   r8 = 0x00000fa86166aeb3   r9 = 0x0000000000000000  r10 = 0x00000f7a58280e85  r11 = 0x0000000000000000\nE   r12 = 0x00000fa86166aeb4  r13 = 0x00007db30b321590  r14 = 0x00007db30b321480  r15 = 0x00007d430b357598\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\nE   Thread T52 (ydbd.System) created by T0 here:\nE   #0 0x0000211121e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149e555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149e555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf50ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 0x000023c8f223 in NActors::TCpuManager::Start() /-S/ydb/library/actors/core/cpu_manager.cpp:139:32\nE   #5 0x000023c4b98d in NActors::TActorSystem::Start() /-S/ydb/library/actors/core/actorsystem.cpp:451:21\nE   #6 0x000043bc3966 in NKikimr::TKikimrRunner::KikimrStart() /-S/ydb/core/driver_lib/run/run.cpp:2214:22\nE   #7 0x000040ac960b in NKikimr::MainRun(NKikimr::TKikimrRunConfig const&, std::__y1::shared_ptr<NKikimr::TModuleFactories>) /-S/ydb/core/driver_lib/run/main.cpp:48:17\nE   #8 0x0000439c6e5b in NKik...","1773677510000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3043, 'EndTime': 1773672394794061, 'Iops': 276, 'RequestsCompleted': 3043, 'StartTime': 1773672383792607, 'WriteLatency': {'Max': 459263, 'Mean': 115038.604, 'Min': 12160, 'P50': 103359, 'P90': 188287, 'P95': 222847, 'P99': 332287, 'P999': 422143, 'StdDeviation': 59254.91152}}","1773676288000000":"Killed by timeout (600 s)","1773675081000000":"Killed by timeout (600 s)","1773674966000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2706, 'EndTime': 1773668782846663, 'Iops': 245, 'RequestsCompleted': 2706, 'StartTime': 1773668771841907, 'WriteLatency': {'Max': 517887, 'Mean': 129437.9675, 'Min': 11352, 'P50': 118079, 'P90': 218879, 'P95': 247807, 'P99': 327679, 'P999': 437759, 'StdDeviation': 64311.52582}}","1773674896000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2551, 'EndTime': 1773669535746865, 'Iops': 231, 'RequestsCompleted': 2551, 'StartTime': 1773669524744942, 'WriteLatency': {'Max': 464895, 'Mean': 136956.7072, 'Min': 21024, 'P50': 124287, 'P90': 217471, 'P95': 255615, 'P99': 339711, 'P999': 414207, 'StdDeviation': 62583.45698}}","1773674709000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2867, 'EndTime': 1773669782507142, 'Iops': 260, 'RequestsCompleted': 2867, 'StartTime': 1773669771504208, 'WriteLatency': {'Max': 532991, 'Mean': 122404.2344, 'Min': 6628, 'P50': 108671, 'P90': 202367, 'P95': 244607, 'P99': 385791, 'P999': 530431, 'StdDeviation': 67805.66206}}","1773674432000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2557, 'EndTime': 1773669318819148, 'Iops': 232, 'RequestsCompleted': 2557, 'StartTime': 1773669307816871, 'WriteLatency': {'Max': 584703, 'Mean': 136481.1873, 'Min': 14408, 'P50': 118271, 'P90': 227071, 'P95': 285951, 'P99': 422911, 'P999': 537087, 'StdDeviation': 75659.56375}}","1773672747000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2840, 'EndTime': 1773667195224752, 'Iops': 258, 'RequestsCompleted': 2840, 'StartTime': 1773667184223956, 'WriteLatency': {'Max': 440575, 'Mean': 122885.7662, 'Min': 19616, 'P50': 112511, 'P90': 198143, 'P95': 237183, 'P99': 318463, 'P999': 410879, 'StdDeviation': 58862.47766}}","1773668553000000":"Killed by timeout (600 s)","1773667038000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2782, 'EndTime': 1773662336818605, 'Iops': 252, 'RequestsCompleted': 2782, 'StartTime': 1773662325816148, 'WriteLatency': {'Max': 783871, 'Mean': 125470.0762, 'Min': 17152, 'P50': 108607, 'P90': 209919, 'P95': 267263, 'P99': 431103, 'P999': 608767, 'StdDeviation': 76234.80306}}","1773661068000000":"Killed by timeout (600 s)","1773661040000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with errorteardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==48655==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x0000507bc257 bp 0x7afd45eeff20 sp 0x7afd45eeff00 T97)\nE   ==48655==The signal is caused by a READ memory access.\nE   ==48655==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x0000507bc257 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x0000507bc257 in unsigned long std::__y1::__hash_table<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::__unordered_map_hasher<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::hash<unsigned long>, std::__y1::equal_to<unsigned long>, true>, std::__y1::__unordered_map_equal<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::equal_to<unsigned long>, std::__y1::hash<unsigned long>, true>, std::__y1::allocator<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>>>::__erase_unique<unsigned long>(unsigned long const&) /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table:1877:18\nE   #2 0x0000507b318c in erase /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1336:80\nE   #3 0x0000507b318c in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:42:26\nE   #4 0x0000507b5070 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x000050783395 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x000050783395 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x000050783395 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x000050783395 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x000050783395 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x000050783395 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507a76eb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507a76eb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507a76eb in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507a744c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x0000507937b3 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x0000507937b3 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x0000507937b3 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x00005079066f in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c312a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d231b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d2cdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d2c37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d2e3ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214aab54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x0000211285a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7efd8ef5cac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7efd8efee8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==48655==Register values:\nE   rax = 0x17d7d7d7d7d7d7d7  rbx = 0x00007b2d97fe2818  rcx = 0x00000000000003e9  rdx = 0x00000f65b2ffc502\nE   rdi = 0xbebebebebebebebe  rsi = 0x0000000000000001  rbp = 0x00007afd45eeff20  rsp = 0x00007afd45eeff00\nE   r8 = 0x0000000000000001   r9 = 0x0000000000000001  r10 = 0x00007bfd8e15f778  r11 = 0x0000000000000002\nE   r12 = 0x00000f6028bd5ff4  r13 = 0x00007bfd8e15f7b0  r14 = 0x00007bfd8e15f758  r15 = 0x00000000000003e9\nE   AddressSanitizer can not provide additional info....","1773659987000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773655409000000":"teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   ======= terminate() call stack ========\nE   Current KQP shutdown state: spent 0 seconds, not started yet\nE   0. /tmp//-S/ydb/core/driver_lib/run/main.cpp:182: KikimrTerminateHandler @ 0x40AB7CE9\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: terminate @ 0x211A7B48\nE   2. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:821: report_failure @ 0x211A8F0E\nE   3. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: __cxa_rethrow @ 0x211A8F0E\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C19517\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D0B1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D14DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D1437D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D163EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x21492B54\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x211105A6\nE   11. ??:0: ?? @ 0x7FA99276FAC2\nE   12. ??:0: ?? @ 0x7FA9928018BF\nE   ======== exception call stack =========\nE   0. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:839: throw_exception @ 0x211A836C\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:882: __cxa_throw @ 0x211A818C\nE   2. /-S/contrib/libs/cxxsupp/libcxx/include/stdexcept:251: ?? @ 0x1DB12D5C\nE   3. /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1787: at @ 0x507580F4\nE   4. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40: ChildSpanEndOk @ 0x507580F4\nE   5. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164: OnWriteFinished @ 0x5075A0E0\nE   6. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365: operator() @ 0x50728405\nE   7. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x50728405\nE   8. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x50728405\nE   9. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x50728405\nE   10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x50728405\nE   11. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x50728405\nE   12. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x5074C75B\nE   13. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x5074C75B\nE   14. /-S/library/cpp/threading/future/core/future-inl.h:210: RunCallbacks @ 0x5074C75B\nE   15. /-S/library/cpp/threading/future/core/future-inl.h:164: TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x5074C4BC\nE   16. /-S/library/cpp/threading/future/core/future-inl.h:136: SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x50738823\nE   17. /-S/library/cpp/threading/future/core/future-inl.h:815: SetValue @ 0x50738823\nE   18. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329: HandleWritePersistentBufferResult @ 0x50738823\nE   19. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653: StateWork @ 0x507356DF\nE   20. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C192A7\nE   21. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:1008: __cxa_rethrow @ 0x211A8DC6\nE   22. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C19517\nE   23. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D0B1B1\nE   24. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D14DC6\nE   25. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D1437D\nE   26. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D163EE\nE   27. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x21492B54\nE   28. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x211105A6\nE   29. ??:0: ?? @ 0x7FA99276FAC2\nE   30. ??:0: ?? @ 0x7FA9928018BF\nE   =======================================\nE   Terminating due to uncaught exception 0x7d49917e3890    what() -> \"unordered_map::at: key not found\"\nE   of type std::out_of_range","1773653670000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2897, 'EndTime': 1773648750728425, 'Iops': 263, 'RequestsCompleted': 2897, 'StartTime': 1773648739727108, 'WriteLatency': {'Max': 327935, 'Mean': 54472.35312, 'Min': 1509, 'P50': 42943, 'P90': 111487, 'P95': 140031, 'P99': 224255, 'P999': 295679, 'StdDeviation': 44844.48496}}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==114409==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x00005081c397 bp 0x7b6dcafb6f20 sp 0x7b6dcafb6f00 T96)\nE   ==114409==The signal is caused by a READ memory access.\nE   ==114409==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   Current KQP shutdown state: spent 0 seconds, not started yet\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x00005081c397 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x00005081c397 in unsigned long std::__y1::__hash_table<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::__unordered_map_hasher<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::hash<unsigned long>, std::__y1::equal_to<unsigned long>, true>, std::__y1::__unordered_map_equal<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::equal_to<unsigned long>, std::__y1::hash<unsigned long>, true>, std::__y1::allocator<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>>>::__erase_unique<unsigned long>(unsigned long const&) /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table:1877:18\nE   #2 0x0000508132cc in erase /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1336:80\nE   #3 0x0000508132cc in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:42:26\nE   #4 0x0000508151b0 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x0000507e34d5 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x0000507e34d5 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x0000507e34d5 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x0000507e34d5 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x0000507e34d5 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x0000507e34d5 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x00005080782b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x00005080782b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x00005080782b in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x00005080758c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x0000507f38f3 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x0000507f38f3 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x0000507f38f3 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x0000507f07af in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c892a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d7b1b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d84dc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d8437d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d863ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x000021503104 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x000021180b56 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7f6e14874ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7f6e149068bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==114409==Register values:\nE   rax = 0x17d7d7d7d7d7d7d7  rbx = 0x00007b9e16b87a48  rcx = 0x000000000000010b  rdx = 0x00000f73c2d70f48\nE   rdi = 0xbebebebebebebebe  rsi = 0x0000000000000001  rbp = 0x00007b6dcafb6f20  rsp = 0x00007b6dcafb6f00\nE   r8 = 0x0000000000000001   r9 = 0x0000000000000001  r10 = 0x00007c6e13905678  r11 = 0x0000000000000002\nE   r12 = 0x00000f6e395eedf4  r13 = 0x00007c6e139056b0  r14 = 0x00007c6e13905658  r15 = 0x000000000000010b\nE  ...","1773651160000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2847, 'EndTime': 1773645928250272, 'Iops': 258, 'RequestsCompleted': 2847, 'StartTime': 1773645917245600, 'WriteLatency': {'Max': 503039, 'Mean': 123115.9101, 'Min': 13480, 'P50': 112063, 'P90': 198527, 'P95': 237055, 'P99': 323071, 'P999': 461823, 'StdDeviation': 60521.06093}}","1773643731000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2606, 'EndTime': 1773638923280013, 'Iops': 236, 'RequestsCompleted': 2606, 'StartTime': 1773638912260247, 'WriteLatency': {'Max': 611327, 'Mean': 134711.2863, 'Min': 16336, 'P50': 122367, 'P90': 212607, 'P95': 252031, 'P99': 348159, 'P999': 518655, 'StdDeviation': 63207.69752}}","1773642105000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2745, 'EndTime': 1773637537804502, 'Iops': 249, 'RequestsCompleted': 2745, 'StartTime': 1773637526802293, 'WriteLatency': {'Max': 576511, 'Mean': 127367.3559, 'Min': 11112, 'P50': 113151, 'P90': 216447, 'P95': 247039, 'P99': 336383, 'P999': 452607, 'StdDeviation': 66072.7907}}","1773634122000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3071, 'EndTime': 1773628564676456, 'Iops': 279, 'RequestsCompleted': 3071, 'StartTime': 1773628553675932, 'WriteLatency': {'Max': 468479, 'Mean': 114239.3852, 'Min': 8768, 'P50': 101119, 'P90': 194815, 'P95': 233215, 'P99': 346623, 'P999': 430335, 'StdDeviation': 64868.24949}}","1773615308000000":"Killed by timeout (600 s)","1773608284000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3052, 'EndTime': 1773602883423493, 'Iops': 277, 'RequestsCompleted': 3052, 'StartTime': 1773602872422671, 'WriteLatency': {'Max': 402687, 'Mean': 114264.6527, 'Min': 2186, 'P50': 101567, 'P90': 196607, 'P95': 231935, 'P99': 294143, 'P999': 384767, 'StdDeviation': 59484.81224}}","1773592886000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2963, 'EndTime': 1773591803329042, 'Iops': 267, 'RequestsCompleted': 2963, 'StartTime': 1773591792254298, 'WriteLatency': {'Max': 413951, 'Mean': 119108.0418, 'Min': 15624, 'P50': 110463, 'P90': 187135, 'P95': 213759, 'P99': 266495, 'P999': 344063, 'StdDeviation': 49590.73091}}","1773572070000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2533, 'EndTime': 1773567013112069, 'Iops': 230, 'RequestsCompleted': 2533, 'StartTime': 1773567002110640, 'WriteLatency': {'Max': 593919, 'Mean': 138181.7955, 'Min': 17920, 'P50': 122879, 'P90': 233855, 'P95': 284159, 'P99': 375039, 'P999': 474111, 'StdDeviation': 72175.90224}}","1773562210000000":"Killed by timeout (600 s)","1773547892000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2920, 'EndTime': 1773542485723158, 'Iops': 265, 'RequestsCompleted': 2920, 'StartTime': 1773542474722529, 'WriteLatency': {'Max': 541695, 'Mean': 119959.0233, 'Min': 11024, 'P50': 107455, 'P90': 203391, 'P95': 244479, 'P99': 324095, 'P999': 447487, 'StdDeviation': 63710.82438}}","1773519197000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksRead': 722, 'EndTime': 1773514432643683, 'Iops': 65, 'ReadLatency': {'Max': 110015, 'Mean': 9442.903047, 'Min': 575, 'P50': 7319, 'P90': 17535, 'P95': 22687, 'P99': 58559, 'P999': 98687, 'StdDeviation': 10177.25145}, 'RequestsCompleted': 722, 'StartTime': 1773514421642360}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==289008==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x000050827307 bp 0x7bb35a60bf20 sp 0x7bb35a60bf00 T97)\nE   ==289008==The signal is caused by a READ memory access.\nE   ==289008==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   Current KQP shutdown state: spent 0 seconds, not started yet\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x000050827307 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x000050827307 in unsigned long std::__y1::__hash_table<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::__unordered_map_hasher<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::hash<unsigned long>, std::__y1::equal_to<unsigned long>, true>, std::__y1::__unordered_map_equal<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::equal_to<unsigned long>, std::__y1::hash<unsigned long>, true>, std::__y1::allocator<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>>>::__erase_unique<unsigned long>(unsigned long const&) /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table:1877:18\nE   #2 0x00005081e23c in erase /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1336:80\nE   #3 0x00005081e23c in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:42:26\nE   #4 0x000050820120 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x0000507ee445 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x0000507ee445 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x0000507ee445 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x0000507ee445 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x0000507ee445 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x0000507ee445 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x00005081279b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x00005081279b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x00005081279b in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000508124fc in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x0000507fe863 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x0000507fe863 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x0000507fe863 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x0000507fb71f in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c8c2a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d7e1b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d87dc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d8737d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d893ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x000021506064 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x000021183ab6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7fb3a3699ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7fb3a372b8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==289008==Register values:\nE   rax = 0x17d7d7d7d7d7d7d7  rbx = 0x00007be3a572b218  rcx = 0x00000000000003c1  rdx = 0x00000f7c74ae5642\nE   rdi = 0xbebebebebebebebe  rsi = 0x0000000000000001  rbp = 0x00007bb35a60bf20  rsp = 0x00007bb35a60bf00\nE   r8 = 0x0000000000000001   r9 = 0x0000000000000001  r10 = 0x00007cb3a2804678  r11 = 0x0000000000000002\nE   r12 = 0x00000f76eb4b97f4  r13 = 0x00007cb3a28046b0  r14 = 0x00007cb3a2804658  r15 = 0x00000000000003c1\nE  ...","1773495175000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3011, 'EndTime': 1773490696917276, 'Iops': 273, 'RequestsCompleted': 3011, 'StartTime': 1773490685914495, 'WriteLatency': {'Max': 421887, 'Mean': 116203.7489, 'Min': 12336, 'P50': 103231, 'P90': 193791, 'P95': 236415, 'P99': 312575, 'P999': 366079, 'StdDeviation': 58052.69789}}","1773488108000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3042, 'EndTime': 1773482861026136, 'Iops': 276, 'RequestsCompleted': 3042, 'StartTime': 1773482850021432, 'WriteLatency': {'Max': 550911, 'Mean': 115066.4957, 'Min': 9096, 'P50': 102591, 'P90': 190591, 'P95': 225535, 'P99': 390655, 'P999': 504319, 'StdDeviation': 65156.90957}}","1773485812000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2430, 'EndTime': 1773480758729742, 'Iops': 220, 'RequestsCompleted': 2430, 'StartTime': 1773480747726297, 'WriteLatency': {'Max': 450815, 'Mean': 143290.6272, 'Min': 21408, 'P50': 131327, 'P90': 227199, 'P95': 271871, 'P99': 373503, 'P999': 444415, 'StdDeviation': 65469.98627}}","1773485425000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3054, 'EndTime': 1773480518946673, 'Iops': 277, 'RequestsCompleted': 3054, 'StartTime': 1773480507946128, 'WriteLatency': {'Max': 460031, 'Mean': 114806.8933, 'Min': 4320, 'P50': 101439, 'P90': 187263, 'P95': 229631, 'P99': 345343, 'P999': 432895, 'StdDeviation': 61474.30575}}","1773481064000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2583, 'EndTime': 1773479837418772, 'Iops': 234, 'RequestsCompleted': 2583, 'StartTime': 1773479826419859, 'WriteLatency': {'Max': 500991, 'Mean': 135250.1711, 'Min': 18848, 'P50': 120959, 'P90': 213503, 'P95': 259327, 'P99': 368127, 'P999': 406271, 'StdDeviation': 63533.68752}}","1773464285000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksRead': 1056, 'EndTime': 1773463791424647, 'Iops': 95, 'ReadLatency': {'Max': 51167, 'Mean': 4117.260417, 'Min': 395, 'P50': 2911, 'P90': 8423, 'P95': 11439, 'P99': 19215, 'P999': 49055, 'StdDeviation': 4232.971925}, 'RequestsCompleted': 1056, 'StartTime': 1773463780424247}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==1831290==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x0000507c1a67 bp 0x7b2e7201ff20 sp 0x7b2e7201ff00 T98)\nE   ==1831290==The signal is caused by a READ memory access.\nE   ==1831290==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x0000507c1a67 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x0000507c1a67 in unsigned long std::__y1::__hash_table<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::__unordered_map_hasher<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::hash<unsigned long>, std::__y1::equal_to<unsigned long>, true>, std::__y1::__unordered_map_equal<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::equal_to<unsigned long>, std::__y1::hash<unsigned long>, true>, std::__y1::allocator<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>>>::__erase_unique<unsigned long>(unsigned long const&) /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table:1877:18\nE   #2 0x0000507b899c in erase /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1336:80\nE   #3 0x0000507b899c in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:42:26\nE   #4 0x0000507ba880 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x000050788ba5 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x000050788ba5 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x000050788ba5 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x000050788ba5 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x000050788ba5 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x000050788ba5 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507acefb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507acefb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507acefb in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507acc5c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x000050798fc3 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x000050798fc3 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x000050798fc3 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x000050795e7f in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c302a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d221b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d2bdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d2b37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d2d3ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214a9ab4 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x000021127506 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7f2ebb7aaac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7f2ebb83c8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==1831290==Register values:\nE   rax = 0x17d7d7d7d7d7d7d7  rbx = 0x00007b5ec45220c8  rcx = 0x0000000000000600  rdx = 0x00000f6bd88a4418\nE   rdi = 0xbebebebebebebebe  rsi = 0x0000000000000000  rbp = 0x00007b2e7201ff20  rsp = 0x00007b2e7201ff00\nE   r8 = 0x0000000000000001   r9 = 0x0000000000000001  r10 = 0x00007c2eba9ec678  r11 = 0x0000000000000002\nE   r12 = 0x00000f664e3fbff4  r13 = 0x00007c2eba9ec6b0  r14 = 0x00007c2eba9ec658  r15 = 0x0000000000000600\nE   AddressSanitizer can not provide addition...","1773463643000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksRead': 1075, 'EndTime': 1773457671033762, 'Iops': 97, 'ReadLatency': {'Max': 86015, 'Mean': 8885.574884, 'Min': 562, 'P50': 6727, 'P90': 17887, 'P95': 24143, 'P99': 44831, 'P999': 76159, 'StdDeviation': 9043.344259}, 'RequestsCompleted': 1075, 'StartTime': 1773457660033238}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==860449==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x0000507b87e7 bp 0x7bf15b5fff50 sp 0x7bf15b5fff30 T98)\nE   ==860449==The signal is caused by a READ memory access.\nE   ==860449==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x0000507b87e7 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x0000507b87e7 in find /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1393:78\nE   #2 0x0000507b87e7 in at /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1785:18\nE   #3 0x0000507b87e7 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40:39\nE   #4 0x0000507ba880 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x000050788ba5 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x000050788ba5 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x000050788ba5 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x000050788ba5 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x000050788ba5 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x000050788ba5 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507acefb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507acefb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507acefb in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507acc5c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x000050798fc3 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x000050798fc3 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x000050798fc3 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x000050795e7f in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c302a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d221b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d2bdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d2b37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d2d3ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214a9ab4 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x000021127506 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7ff1a4e5fac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7ff1a4ef18bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==860449==Register values:\nE   rax = 0xbebebebebebebebe  rbx = 0x00000f9e34800b56  rcx = 0x0000000000000001  rdx = 0x17d7d7d7d7d7d7d7\nE   rdi = 0x00007cf1a4005a78  rsi = 0x0000000000000309  rbp = 0x00007bf15b5fff50  rsp = 0x00007bf15b5fff30\nE   r8 = 0x0000000000000002   r9 = 0x0000000000000001  r10 = 0x00000f7e2b6bfff3  r11 = 0x0000000000000000\nE   r12 = 0x00000f7eab6b7ff4  r13 = 0x00007cf1a4005ab0  r14 = 0x00007cf1a4005a58  r15 = 0x0000000000000309\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table in find<unsigned long>\nE   Thread T98 (ydbd.System) created by T0 here:\nE   #0 0x00002110e141 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149a4b5 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149a4b5 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf10ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 0x000023c8b223 in NActors::TCpuManager::Start() /-S/ydb/libra...","1773428652000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2683, 'EndTime': 1773423676700881, 'Iops': 243, 'RequestsCompleted': 2683, 'StartTime': 1773423665698872, 'WriteLatency': {'Max': 557055, 'Mean': 130569.801, 'Min': 22432, 'P50': 118463, 'P90': 217855, 'P95': 248319, 'P99': 331775, 'P999': 455935, 'StdDeviation': 63138.48558}}","1773422300000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","1773418484000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2656, 'EndTime': 1773413199563054, 'Iops': 241, 'RequestsCompleted': 2656, 'StartTime': 1773413188559521, 'WriteLatency': {'Max': 731135, 'Mean': 131936.4409, 'Min': 3722, 'P50': 116031, 'P90': 217855, 'P95': 257407, 'P99': 406527, 'P999': 598015, 'StdDeviation': 74241.13829}}","1773418209000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2940, 'EndTime': 1773413120221082, 'Iops': 267, 'RequestsCompleted': 2940, 'StartTime': 1773413109217185, 'WriteLatency': {'Max': 516607, 'Mean': 118782.8163, 'Min': 11392, 'P50': 104383, 'P90': 207359, 'P95': 247295, 'P99': 326911, 'P999': 463359, 'StdDeviation': 65739.37718}}","1773416973000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksRead': 844, 'EndTime': 1773412172878794, 'Iops': 76, 'ReadLatency': {'Max': 96447, 'Mean': 12035.74408, 'Min': 497, 'P50': 8895, 'P90': 27631, 'P95': 33791, 'P99': 47903, 'P999': 71231, 'StdDeviation': 11634.42692}, 'RequestsCompleted': 844, 'StartTime': 1773412161878007}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==157197==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x0000507bbe27 bp 0x7b5e0a4d7f50 sp 0x7b5e0a4d7f30 T98)\nE   ==157197==The signal is caused by a READ memory access.\nE   ==157197==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   Current KQP shutdown state: spent 0 seconds, not started yet\nE   #0 0x0000507bbe27 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x0000507bbe27 in find /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1393:78\nE   #2 0x0000507bbe27 in at /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1785:18\nE   #3 0x0000507bbe27 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40:39\nE   #4 0x0000507bdec0 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x00005078c1e5 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x00005078c1e5 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x00005078c1e5 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x00005078c1e5 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x00005078c1e5 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x00005078c1e5 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507b053b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507b053b in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507b053b in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507b029c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x00005079c603 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x00005079c603 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x00005079c603 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x0000507994bf in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c302a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d221b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d2bdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d2b37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d2d3ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214a9ab4 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x000021127506 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7f5e53d1fac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7f5e53db18bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==157197==Register values:\nE   rax = 0xbebebebebebebebe  rbx = 0x00000f8bca5def16  rcx = 0x0000000000000001  rdx = 0x17d7d7d7d7d7d7d7\nE   rdi = 0x00007c5e52ef7878  rsi = 0x000000000000034d  rbp = 0x00007b5e0a4d7f50  rsp = 0x00007b5e0a4d7f30\nE   r8 = 0x0000000000000002   r9 = 0x0000000000000001  r10 = 0x00000f6bc149aff3  r11 = 0x0000000000000000\nE   r12 = 0x00000f6c41492ff4  r13 = 0x00007c5e52ef78b0  r14 = 0x00007c5e52ef7858  r15 = 0x000000000000034d\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table in find<unsigned long>\nE   Thread T98 (ydbd.System) created by T0 here:\nE   #0 0x00002110e141 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149a4b5 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149a4b5 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf10ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 0x000023c8b223 ...","1773416491000000":"Killed by timeout (600 s)","1773412278000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3004, 'EndTime': 1773406893567748, 'Iops': 272, 'RequestsCompleted': 3004, 'StartTime': 1773406882562022, 'WriteLatency': {'Max': 526847, 'Mean': 116349.7197, 'Min': 5916, 'P50': 103871, 'P90': 194431, 'P95': 231551, 'P99': 307455, 'P999': 474623, 'StdDeviation': 61056.84668}}","1773408162000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3005, 'EndTime': 1773403057534307, 'Iops': 273, 'RequestsCompleted': 3005, 'StartTime': 1773403046533075, 'WriteLatency': {'Max': 625663, 'Mean': 116465.5907, 'Min': 14744, 'P50': 105023, 'P90': 194175, 'P95': 231935, 'P99': 321535, 'P999': 471807, 'StdDeviation': 59886.85769}}","1773407266000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:72: in test_nbs_load_actor_write_then_read\n    actor_id = self.get_load_actor_adapter_actor_id(disk_id)\nydb/tests/functional/nbs/common.py:130: in get_load_actor_adapter_actor_id\n    assert status == \"success\"\nE   AssertionErrorteardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_3/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_3/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T12:48:25.997835Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214C5CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214B355A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421FB1BE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421A8941\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C512A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D431B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D4CDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4C37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D4E3EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214CAAB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21148506\nE   11. ??:0: ?? @ 0x7FB135453AC2\nE   12. ??:0: ?? @ 0x7FB1354E58BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_4/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_4/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T12:48:25.997862Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214C5CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214B355A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421FB1BE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421A8941\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C512A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D431B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D4CDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4C37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D4E3EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214CAAB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21148506\nE   11. ??:0: ?? @ 0x7F41AB592AC2\nE   12. ??:0: ?? @ 0x7F41AB6248BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_5/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_5/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T12:48:26.012619Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214C5CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214B355A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421FB1BE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421A8941\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C512A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D431B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D4CDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4C37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D4E3EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214CAAB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21148506\nE   11. ??:0: ?? @ 0x7FE417436AC2\nE   12. ??:0: ?? @ 0x7FE4174C88BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_6/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_6/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T12:48:25.994855Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214C5CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214B355A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421FB1BE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x421A8941\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C512A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D431B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D4CDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4C37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D4E3EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214CAAB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21148506\nE   11. ??:0: ?? @ 0x7FC46974BAC2\nE   12. ??:0: ?? @ 0x7FC4697DD8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_7/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py...","1773404601000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2364, 'EndTime': 1773400023468590, 'Iops': 214, 'RequestsCompleted': 2364, 'StartTime': 1773400012468459, 'WriteLatency': {'Max': 641535, 'Mean': 147909.269, 'Min': 19952, 'P50': 131967, 'P90': 237439, 'P95': 287999, 'P99': 394239, 'P999': 624127, 'StdDeviation': 71905.37763}}","1773398367000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3046, 'EndTime': 1773396832605779, 'Iops': 276, 'RequestsCompleted': 3046, 'StartTime': 1773396821604319, 'WriteLatency': {'Max': 482303, 'Mean': 89938.77479, 'Min': 2666, 'P50': 79871, 'P90': 161407, 'P95': 198015, 'P99': 280063, 'P999': 404223, 'StdDeviation': 56563.14213}}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Bad exit_code..\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   ======= terminate() call stack ========\nE   Current KQP shutdown state: spent 1e-06 seconds, not started yet\nE   0. /tmp//-S/ydb/core/driver_lib/run/main.cpp:182: KikimrTerminateHandler @ 0x40AEED79\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: terminate @ 0x211D9AA8\nE   2. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:821: report_failure @ 0x211DAE6E\nE   3. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:0: __cxa_rethrow @ 0x211DAE6E\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C4B517\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3D1B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D46DC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4637D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D483EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C4AB4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21142506\nE   11. ??:0: ?? @ 0x7FA95A54DAC2\nE   12. ??:0: ?? @ 0x7FA95A5DF8BF\nE   ======== exception call stack =========\nE   0. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:839: throw_exception @ 0x211DA2CC\nE   1. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:882: __cxa_throw @ 0x211DA0EC\nE   2. /-S/contrib/libs/cxxsupp/libcxx/include/stdexcept:251: ?? @ 0x1DB44D5C\nE   3. /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1787: at @ 0x507B11B4\nE   4. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40: ChildSpanEndOk @ 0x507B11B4\nE   5. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164: OnWriteFinished @ 0x507B31A0\nE   6. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365: operator() @ 0x507814C5\nE   7. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x507814C5\nE   8. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x507814C5\nE   9. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> @ 0x507814C5\nE   10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x507814C5\nE   11. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x507814C5\nE   12. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x507A581B\nE   13. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x507A581B\nE   14. /-S/library/cpp/threading/future/core/future-inl.h:210: RunCallbacks @ 0x507A581B\nE   15. /-S/library/cpp/threading/future/core/future-inl.h:164: TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x507A557C\nE   16. /-S/library/cpp/threading/future/core/future-inl.h:136: SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> @ 0x507918E3\nE   17. /-S/library/cpp/threading/future/core/future-inl.h:815: SetValue @ 0x507918E3\nE   18. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329: HandleWritePersistentBufferResult @ 0x507918E3\nE   19. /tmp//-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653: StateWork @ 0x5078E79F\nE   20. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C4B2A7\nE   21. /tmp//-S/contrib/libs/cxxsupp/libcxxrt/exception.cc:1008: __cxa_rethrow @ 0x211DAD26\nE   22. /tmp//-S/ydb/library/actors/core/actor.cpp:358: Receive @ 0x23C4B517\nE   23. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D3D1B1\nE   24. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D46DC6\nE   25. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D4637D\nE   26. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D483EE\nE   27. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214C4AB4\nE   28. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21142506\nE   29. ??:0: ?? @ 0x7FA95A54DAC2\nE   30. ??:0: ?? @ 0x7FA95A5DF8BF\nE   =======================================\nE   Terminating due to uncaught exception 0x7d495950c490    what() -> \"unordered_map::at: key not found\"\nE   of type std::out_of_range","1773397572000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2613, 'EndTime': 1773392732995337, 'Iops': 237, 'RequestsCompleted': 2613, 'StartTime': 1773392721995044, 'WriteLatency': {'Max': 516351, 'Mean': 133731.3219, 'Min': 18384, 'P50': 117055, 'P90': 226559, 'P95': 272127, 'P99': 346367, 'P999': 446207, 'StdDeviation': 67843.58745}}","1773397083000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with errorteardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==947361==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x0000508072a7 bp 0x7bbe3251ef50 sp 0x7bbe3251ef30 T97)\nE   ==947361==The signal is caused by a READ memory access.\nE   ==947361==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x0000508072a7 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x0000508072a7 in find /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1393:78\nE   #2 0x0000508072a7 in at /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1785:18\nE   #3 0x0000508072a7 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:40:39\nE   #4 0x000050809340 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x0000507d7645 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x0000507d7645 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x0000507d7645 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x0000507d7645 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x0000507d7645 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x0000507d7645 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507fb9bb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507fb9bb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507fb9bb in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507fb71c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x0000507e7a83 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x0000507e7a83 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x0000507e7a83 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x0000507e493f in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023ca22a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d941b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d9ddc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d9d37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d9f3ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x00002151c064 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x000021199ab6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7fbe7b5a9ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7fbe7b63b8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==947361==Register values:\nE   rax = 0xbebebebebebebebe  rbx = 0x00000f97cf4d4e56  rcx = 0x0000000000000001  rdx = 0x17d7d7d7d7d7d7d7\nE   rdi = 0x00007cbe7a6a7278  rsi = 0x00000000000001cf  rbp = 0x00007bbe3251ef50  rsp = 0x00007bbe3251ef30\nE   r8 = 0x0000000000000002   r9 = 0x0000000000000001  r10 = 0x00000f77c64a3df3  r11 = 0x0000000000000000\nE   r12 = 0x00000f784649bdf4  r13 = 0x00007cbe7a6a72b0  r14 = 0x00007cbe7a6a7258  r15 = 0x00000000000001cf\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table in find<unsigned long>\nE   Thread T97 (ydbd.System) created by T0 here:\nE   #0 0x0000211806f1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002150ca65 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002150ca65 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023d630ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 0x000023cfd223 in NActors::TCpuManager::Start() /-S/ydb/libra...","1773391749000000":"Killed by timeout (600 s)","1773389368000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:71: in test_nbs_load_actor_write_then_read\n    self.create_disk(disk_id)\nydb/tests/functional/nbs/common.py:102: in create_disk\n    execute_dstool_grpc(\nydb/tests/functional/nbs/helpers.py:38: in execute_dstool_grpc\n    False\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:24572', 'nbs', 'partition', 'create', '--pool', 'ddp1', '--block-size=4096', '--blocks-count=1048576', '--type=ssd', '--disk-id', 'R0dLtWAb5Y']\nE    finished with exit code 1, stderr:\nE   \nE   WARNING: failed to fetch data from host localhost:24572 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:24572 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:24572 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:24572 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   WARNING: failed to fetch data from host localhost:24572 in invoke_grpc: Can't connect to specified addresses by gRPC protocol (ydb.apps.dstool.lib.common.ConnectionError)\nE   Connection Error: Can't connect to specified addresses\nE   \nE   \nE   stdout:teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T07:46:18.560064Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214B7E07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A569A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421E8BFE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x42196381\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C432A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D351B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D3EDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3E37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D403EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BCBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113B506\nE   11. ??:0: ?? @ 0x7F2399998AC2\nE   12. ??:0: ?? @ 0x7F2399A2A8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_2/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_2/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T07:46:18.595330Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214B7E07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A569A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421E8BFE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x42196381\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C432A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D351B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D3EDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3E37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D403EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BCBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113B506\nE   11. ??:0: ?? @ 0x7FD37DC1FAC2\nE   12. ??:0: ?? @ 0x7FD37DCB18BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_8/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_8/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T07:46:18.581871Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214B7E07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A569A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421E8BFE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x42196381\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C432A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D351B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D3EDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3E37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D403EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BCBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113B506\nE   11. ??:0: ?? @ 0x7F07CF128AC2\nE   12. ??:0: ?? @ 0x7F07CF1BA8BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_9/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_9/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T07:46:18.632709Z):\nE   ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204\nE   Handle(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214B7E07\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x214A569A\nE   2. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor_boot.cpp:204: Handle @ 0x421E8BFE\nE   3. /tmp//-S/ydb/core/blobstorage/ddisk/ddisk_actor.cpp:129: StateFunc @ 0x42196381\nE   4. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C432A7\nE   5. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D351B1\nE   6. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D3EDC6\nE   7. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D3E37D\nE   8. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D403EE\nE   9. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214BCBF4\nE   10. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x2113B506\nE   11. ??:0: ?? @ 0x7F7942684AC2\nE   12. ??:0: ?? @ 0x7F79427168BF\nE   Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/node_3/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py...","1773389248000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 2788, 'EndTime': 1773384475168940, 'Iops': 253, 'RequestsCompleted': 2788, 'StartTime': 1773384464169464, 'WriteLatency': {'Max': 455679, 'Mean': 125084.8422, 'Min': 8368, 'P50': 112063, 'P90': 204927, 'P95': 249087, 'P99': 327935, 'P999': 372735, 'StdDeviation': 59842.17618}}","1773366191000000":"ydb/tests/functional/nbs/test_nbs_load_actor.py:84: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(read_results, expected_blocks_read=True, expected_blocks_written=False)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksRead': 958, 'EndTime': 1773364801742558, 'Iops': 87, 'ReadLatency': {'Max': 75327, 'Mean': 6713.823591, 'Min': 615, 'P50': 4651, 'P90': 13639, 'P95': 20127, 'P99': 36383, 'P999': 56895, 'StdDeviation': 7441.262026}, 'RequestsCompleted': 958, 'StartTime': 1773364790741855}teardown failed:\nydb/tests/functional/nbs/common.py:50: in setup\n    self.cluster.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/nbs/test-results/py3test/testing_out_stuff/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==1100623==ERROR: AddressSanitizer: SEGV on unknown address (pc 0x00005080a567 bp 0x7bd53cedcf60 sp 0x7bd53cedcf40 T98)\nE   ==1100623==The signal is caused by a READ memory access.\nE   ==1100623==Hint: this fault was caused by a dereference of a high value address (see register values below).  Disassemble the provided pc to learn which register was used.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   #0 0x00005080a567 in find<unsigned long> /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table\nE   #1 0x00005080a567 in unsigned long std::__y1::__hash_table<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::__unordered_map_hasher<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::hash<unsigned long>, std::__y1::equal_to<unsigned long>, true>, std::__y1::__unordered_map_equal<unsigned long, std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>, std::__y1::equal_to<unsigned long>, std::__y1::hash<unsigned long>, true>, std::__y1::allocator<std::__y1::__hash_value_type<unsigned long, NWilson::TSpan>>>::__erase_unique<unsigned long>(unsigned long const&) /-S/contrib/libs/cxxsupp/libcxx/include/__hash_table:1877:18\nE   #2 0x00005080149c in erase /-S/contrib/libs/cxxsupp/libcxx/include/unordered_map:1336:80\nE   #3 0x00005080149c in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TBaseRequestHandler::ChildSpanEndOk(unsigned long) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:42:26\nE   #4 0x000050803380 in NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler::OnWriteFinished(unsigned long, NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/request.cpp:164:9\nE   #5 0x0000507d1685 in operator() /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:365:33\nE   #6 0x0000507d1685 in __invoke<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179:25\nE   #7 0x0000507d1685 in __call<(lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251:5\nE   #8 0x0000507d1685 in __invoke_r<void, (lambda at /-S/ydb/core/nbs/cloud/blockstore/libs/storage/partition_direct/direct_block_group.cpp:363:13) &, const NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> &> /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273:10\nE   #9 0x0000507d1685 in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167:12\nE   #10 0x0000507d1685 in std::__y1::__function::__func<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0, std::__y1::allocator<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TDirectBlockGroup::DoWriteBlocksLocal(std::__y1::shared_ptr<NYdb::NBS::NBlockStore::NStorage::NPartitionDirect::TWriteRequestHandler>)::$_0>, void (NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&)>::operator()(NThreading::TFuture<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> const&) /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319:10\nE   #11 0x0000507f59fb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436:12\nE   #12 0x0000507f59fb in operator() /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995:10\nE   #13 0x0000507f59fb in NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::RunCallbacks() /-S/library/cpp/threading/future/core/future-inl.h:210:25\nE   #14 0x0000507f575c in bool NThreading::NImpl::TFutureState<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>::TrySetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult>(NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult&&, bool) /-S/library/cpp/threading/future/core/future-inl.h:164:21\nE   #15 0x0000507e1ac3 in SetValue<NKikimrBlobStorage::NDDisk::TEvWritePersistentBufferResult> /-S/library/cpp/threading/future/core/future-inl.h:136:32\nE   #16 0x0000507e1ac3 in SetValue /-S/library/cpp/threading/future/core/future-inl.h:815:16\nE   #17 0x0000507e1ac3 in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::HandleWritePersistentBufferResult(TAutoPtr<NActors::TEventHandle<NKikimr::NDDisk::TEvWritePersistentBufferResult>, TDelete> const&, NActors::TActorContext const&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:329:33\nE   #18 0x0000507de97f in NYdb::NBS::NBlockStore::NStorage::NTransport::TICStorageTransportActor::StateWork(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/core/nbs/cloud/blockstore/libs/storage/storage_transport/ic_storage_transport.cpp:653:9\nE   #19 0x000023c672a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #20 0x000023d591b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #21 0x000023d62dc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #22 0x000023d6237d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #23 0x000023d643ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #24 0x0000214e0bf4 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #25 0x00002115f506 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #26 0x7fd58676fac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #27 0x7fd5868018bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==1100623==Register values:\nE   rax = 0x17d7d7d7d7d7d7d7  rbx = 0x00007c05876c8808  rcx = 0x000000000000055c  rdx = 0x00000f80b0ed9100\nE   rdi = 0xbebebebebebebebe  rsi = 0x0000000000000000  rbp = 0x00007bd53cedcf60  rsp = 0x00007bd53cedcf40\nE   r8 = 0x0000000000000001   r9 = 0x0000000000000001  r10 = 0x00007cd585845178  r11 = 0x0000000000000002\nE   r12 = 0x00000f7b279d39fc  r13 = 0x00007cd5858451b0  r14 = 0x00007cd585845158  r15 = 0x000000000000055c\nE   AddressSanitizer can not provide addition..."},"ydb/tests/functional/statistics/test_restarts.py.test_basic":{"1773735610000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773729526000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773723207000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773722559000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773720290000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773719674000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773717082000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773708708000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773702149000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773702129000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773698398000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773694511000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773692039000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773690853000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773690441000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773690098000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773689559000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773689025000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773688476000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773687448000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773687172000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773687167000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773687158000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773687146000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773685622000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773685521000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773683098000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773682225000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773682040000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773681818000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773680250000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773679933000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773679449000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773679120000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773679007000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773678560000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:621: in __run\n    self.client.bind_storage_pools(self.domain_name, pools, token=root_token)\nydb/tests/library/clients/kikimr_client.py:197: in bind_storage_pools\n    self.invoke(\nydb/tests/library/clients/kikimr_client.py:122: in invoke\n    return callee(request, timeout=self.__timeout)\ncontrib/python/grpcio/py3/grpc/_channel.py:1181: in __call__\n    return _end_unary_response_blocking(state, call, False, None)\ncontrib/python/grpcio/py3/grpc/_channel.py:1006: in _end_unary_response_blocking\n    raise _InactiveRpcError(state)  # pytype: disable=not-instantiable\nE   grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:\nE   \tstatus = StatusCode.UNAVAILABLE\nE   \tdetails = \"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:14791: Failed to connect to remote host: Connection refused\"\nE   \tdebug_error_string = \"UNKNOWN:Error received from peer  {created_time:\"2026-03-16T15:40:10.757872923+00:00\", grpc_status:14, grpc_message:\"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:14791: Failed to connect to remote host: Connection refused\"}\"\nE   >\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/library/fixtures/__init__.py:54: in ydb_cluster\n    cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/statistics/test-results/py3test/testing_out_stuff/test_restarts.py.test_basic/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/statistics/test-results/py3test/testing_out_stuff/test_restarts.py.test_basic/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==560865==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7b20972e1f40 sp 0x7b20972e1e70 T52)\nE   ==560865==The signal is caused by a WRITE memory access.\nE   ==560865==Hint: address points to the zero page.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   AddressSanitizer:DEADLYSIGNAL\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7f20e021cac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7f20e02ae8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==560865==Register values:\nE   rax = 0x00007c90df146110  rbx = 0x00007c90df146120  rcx = 0x0000000000000000  rdx = 0x00000f921be28c22\nE   rdi = 0x00007d00df1412a8  rsi = 0x00000fa01be28255  rbp = 0x00007b20972e1f40  rsp = 0x00007b20972e1e70\nE   r8 = 0x00000f921be28c23   r9 = 0x0000000000000000  r10 = 0x00000f6412a46e85  r11 = 0x0000000000000000\nE   r12 = 0x00000f921be28c24  r13 = 0x00007d00df140d90  r14 = 0x00007d00df140c80  r15 = 0x00007c90df146118\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\nE   Thread T52 (ydbd.System) created by T0 here:\nE   #0 0x0000211121e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149e555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149e555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf50ec in NActors::TBasicExecutorPool::Start() /-S/ydb/library/actors/core/executor_pool_basic.cpp:598:32\nE   #4 ...","1773678224000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773677952000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773677718000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773677510000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773677230000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773677073000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773676288000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773675427000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773675081000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773674966000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773674896000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773674709000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773674432000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773672978000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773672747000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773671187000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773668553000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773668400000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773667338000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773667038000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773666791000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773665446000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773663833000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773661068000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773660879000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773659987000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773659879000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773655409000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773653768000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773653670000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773651160000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773644268000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773643731000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773642105000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773636520000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773634122000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773627098000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773626829000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773622583000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773617923000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773615765000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773615308000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773608284000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773606561000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773592886000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773572070000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773570977000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773562367000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773562210000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773549981000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773547892000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773526956000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773519197000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773517997000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773510184000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773508856000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773495254000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773495175000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773494895000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773489674000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773488352000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773488108000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773485812000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773485425000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773485307000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773485102000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773481064000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773463643000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773428652000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773427126000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773423092000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773422300000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773421818000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773421534000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773419832000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773419396000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773419081000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773418484000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773418466000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773418209000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773417840000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773417667000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773417296000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773417175000000":"ydb/tests/functional/statistics/test_restarts.py:33: in test_basic\n    session_pool.execute_with_retries(f'''\ncontrib/python/ydb/py3/ydb/query/pool.py:259: in execute_with_retries\n    return retry_operation_sync(wrapped_callee, retry_settings)\ncontrib/python/ydb/py3/ydb/retries.py:163: in retry_operation_sync\n    for next_opt in opt_generator:\ncontrib/python/ydb/py3/ydb/retries.py:153: in retry_operation_impl\n    raise status\ncontrib/python/ydb/py3/ydb/retries.py:118: in retry_operation_impl\n    result = YdbRetryOperationFinalResult(callee(*args, **kwargs))\ncontrib/python/ydb/py3/ydb/query/pool.py:255: in wrapped_callee\n    with self.checkout(timeout=retry_settings.max_session_acquire_timeout) as session:\ncontrib/python/ydb/py3/ydb/query/pool.py:340: in __enter__\n    self._session = self._pool.acquire(self._timeout)\ncontrib/python/ydb/py3/ydb/query/pool.py:117: in acquire\n    session = self._create_new_session(time_left)\ncontrib/python/ydb/py3/ydb/query/pool.py:69: in _create_new_session\n    session.create(settings=BaseRequestSettings().with_timeout(timeout))\ncontrib/python/ydb/py3/ydb/query/session.py:371: in create\n    self._create_call(settings=settings)\ncontrib/python/ydb/py3/ydb/query/session.py:160: in _create_call\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:469: in __call__\n    connection = self._store.get(preferred_endpoint)\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:153: in get\n    raise issues.ConnectionLost(\"Couldn't find valid connection\")\nE   ydb.issues.ConnectionLost: Couldn't find valid connectionteardown failed:\nydb/tests/library/fixtures/__init__.py:117: in ydb_database\n    yield from _ydb_database(ydb_cluster, ydb_root, ydb_safe_test_name)\nydb/tests/library/fixtures/__init__.py:111: in _ydb_database\n    with ydb_database_ctx(cluster, database):\ncontrib/tools/python3/Lib/contextlib.py:148: in __exit__\n    next(self.gen)\nydb/tests/library/fixtures/__init__.py:103: in ydb_database_ctx\n    ydb_cluster.unregister_and_stop_slots(database_nodes)\nydb/tests/library/harness/kikimr_runner.py:726: in unregister_and_stop_slots\n    i.stop()\nydb/tests/library/harness/kikimr_runner.py:303: in stop\n    super(KiKiMRNode, self).stop()\nydb/tests/library/harness/daemon.py:194: in stop\n    if not self.__check_can_launch_stop(\"stop\"):\nydb/tests/library/harness/daemon.py:171: in __check_can_launch_stop\n    raise DaemonError(\nE   ydb.tests.library.harness.daemon.DaemonError: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = -6.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/statistics/test-results/py3test/testing_out_stuff/test_restarts.py.test_basic/cluster/slot_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/functional/statistics/test-results/py3test/testing_out_stuff/test_restarts.py.test_basic/cluster/slot_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   VERIFY failed (2026-03-13T14:46:35.685767Z): verification=IndexConstructor.DeserializeFromProto(indexSchema);fline=update.cpp:24;incorrect_proto=Name: \"__minmax_key\"\nE   MinMaxIndex {\nE   ColumnName: \"key\"\nE   }\nE   ;\nE   ydb/library/actors/core/log.cpp:857\nE   ~TVerifyFormattedRecordWriter(): requirement false failed\nE   0. /-S/util/system/yassert.cpp:86: InternalPanicImpl @ 0x214A5CC7\nE   1. /-S/util/system/yassert.cpp:55: Panic @ 0x2149355A\nE   2. /tmp//-S/ydb/library/actors/core/log.cpp:857: ~TVerifyFormattedRecordWriter @ 0x23D4A6BD\nE   3. /tmp//-S/ydb/core/tx/schemeshard/olap/indexes/update.cpp:24: DeserializeFromProto @ 0x2BA10B2B\nE   4. /tmp//-S/ydb/core/tx/schemeshard/olap/indexes/update.cpp:38: Parse @ 0x2BA11E35\nE   5. /tmp//-S/ydb/core/tx/schemeshard/olap/schema/schema.cpp:106: AddDefaultMinMaxIndexes @ 0x2BA145D8\nE   6. /tmp//-S/ydb/core/tx/schemeshard/olap/operations/create_table.cpp:198: DoDeserialize @ 0x2B9D7206\nE   7. /tmp//-S/ydb/core/tx/schemeshard/olap/operations/create_table.cpp:33: Deserialize @ 0x2B9CC7F1\nE   8. /tmp//-S/ydb/core/tx/schemeshard/olap/operations/create_table.cpp:55: BuildTableInfo @ 0x2B9CC7F1\nE   9. /tmp//-S/ydb/core/tx/schemeshard/olap/operations/create_table.cpp:711: Propose @ 0x2B9C5BBD\nE   10. /tmp//-S/ydb/core/tx/schemeshard/schemeshard__operation.cpp:117: ProcessOperationParts @ 0x2B8A7E66\nE   11. /tmp//-S/ydb/core/tx/schemeshard/schemeshard__operation.cpp:292: IgniteOperation @ 0x2B8AE2F7\nE   12. /tmp//-S/ydb/core/tx/schemeshard/schemeshard__operation.cpp:413: Execute @ 0x2B948E78\nE   13. /tmp//-S/ydb/core/tablet_flat/flat_executor.cpp:2036: ExecuteTransaction @ 0x28F4F3B0\nE   14. /tmp//-S/ydb/core/tablet_flat/flat_executor.cpp:1950: DoExecute @ 0x28F4970A\nE   15. /tmp//-S/ydb/core/tablet_flat/flat_executor.cpp:1964: Execute @ 0x28F51FDE\nE   16. /tmp//-S/ydb/core/tablet_flat/tablet_flat_executed.cpp:62: Execute @ 0x28EE498A\nE   17. /tmp//-S/ydb/core/tablet_flat/tablet_flat_executed.cpp:57: Execute @ 0x28EE498A\nE   18. /tmp//-S/ydb/core/tx/schemeshard/schemeshard_impl.cpp:6160: Handle @ 0x2C92B319\nE   19. /tmp//-S/ydb/core/tx/schemeshard/schemeshard_impl.cpp:5326: StateWork @ 0x2C86BFF8\nE   20. /tmp//-S/ydb/library/actors/core/actor.cpp:354: Receive @ 0x23C312A7\nE   21. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:267: Execute @ 0x23D231B1\nE   22. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:455: operator() @ 0x23D2CDC6\nE   23. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:507: ProcessExecutorPool @ 0x23D2C37D\nE   24. /tmp//-S/ydb/library/actors/core/executor_thread.cpp:533: ThreadProc @ 0x23D2E3EE\nE   25. /-S/util/system/thread.cpp:245: ThreadProxy @ 0x214AAAB4\nE   26. /tmp//-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239: asan_thread_start @ 0x21128506\nE   27. ??:0: ?? @ 0x7F3B3AAA5AC2\nE   28. ??:0: ?? @ 0x7F3B3AB378BF","1773417033000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773416973000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773416913000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773416723000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773416600000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773416491000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773414132000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773413019000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773412901000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773412877000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773412278000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773409896000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773408162000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773407266000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773407147000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773406369000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773405522000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773405071000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773404601000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773402640000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773401100000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773398367000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773397572000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773397083000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773394916000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773394033000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773392201000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773391749000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773391647000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773389368000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773389248000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773389101000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773388600000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773378027000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773377371000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773376892000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773374483000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773366191000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773365943000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773364460000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","1773361816000000":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart"},"ydb/tests/tools/pq_read/test/test_timeout.py.TestTimeout.test_timeout":{"1773719674000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:3537', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:3537', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773717082000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:28310', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:28310', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773692039000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:8323', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:8323', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773690853000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:10778', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:10778', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773689025000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:29260', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:29260', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773687172000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:64584', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:64584', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773679449000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:22251', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:22251', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773677952000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:10531', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:10531', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773663833000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:6726', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:6726', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773642105000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:22519', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:22519', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773617923000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:23817', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:23817', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773608284000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:13989', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:13989', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773592886000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:27288', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:27288', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773572070000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:7279', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:7279', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773562367000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:17610', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:17610', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773526956000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:18799', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:18799', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773485425000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:25696', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:25696', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773485102000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:64906', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:64906', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773481064000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:24717', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:24717', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773427126000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:22995', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:22995', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773421818000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:27979', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:27979', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773418484000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:5396', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:5396', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773416973000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:18497', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:18497', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773412901000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:13852', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:13852', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","1773412877000000":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:22461', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:22461', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})"},"ydb/tests/datashard/ttl/test_ttl.py.TestTTL.test_ttl[table_Date_1_UNIQUE_SYNC-pk_types35-all_types35-index35-Date-UNIQUE-SYNC]":{"1773722559000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773720290000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 5 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773717082000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773690098000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773685622000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773683098000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773679933000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773679007000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:621: in __run\n    self.client.bind_storage_pools(self.domain_name, pools, token=root_token)\nydb/tests/library/clients/kikimr_client.py:197: in bind_storage_pools\n    self.invoke(\nydb/tests/library/clients/kikimr_client.py:122: in invoke\n    return callee(request, timeout=self.__timeout)\ncontrib/python/grpcio/py3/grpc/_channel.py:1181: in __call__\n    return _end_unary_response_blocking(state, call, False, None)\ncontrib/python/grpcio/py3/grpc/_channel.py:1006: in _end_unary_response_blocking\n    raise _InactiveRpcError(state)  # pytype: disable=not-instantiable\nE   grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:\nE   \tstatus = StatusCode.UNAVAILABLE\nE   \tdetails = \"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:13411: Failed to connect to remote host: Connection refused\"\nE   \tdebug_error_string = \"UNKNOWN:Error received from peer  {grpc_message:\"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:13411: Failed to connect to remote host: Connection refused\", grpc_status:14, created_time:\"2026-03-16T15:42:06.030637085+00:00\"}\"\nE   >\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/sql/lib/test_base.py:41: in setup_class\n    cls.cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/datashard/ttl/test-results/py3test/testing_out_stuff/test_ttl.py.TestTTL.test_ttl.table_Date_0__SYNC-pk_types30-all_types30-index30-Date--SYNC/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/datashard/ttl/test-results/py3test/testing_out_stuff/test_ttl.py.TestTTL.test_ttl.table_Date_0__SYNC-pk_types30-all_types30-index30-Date--SYNC/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==568985==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7bc874766f40 sp 0x7bc874766e70 T51)\nE   ==568985==The signal is caused by a WRITE memory access.\nE   ==568985==Hint: address points to the zero page.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   AddressSanitizer:DEADLYSIGNAL\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7fc8bcf7dac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7fc8bd00f8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==568985==Register values:\nE   rax = 0x00007d38bbe51010  rbx = 0x00007d38bbe51020  rcx = 0x0000000000000000  rdx = 0x00000fa7177ca202\nE   rdi = 0x00007da8bbe3faa8  rsi = 0x00000fb5177c7f55  rbp = 0x00007bc874766f40  rsp = 0x00007bc874766e70\nE   r8 = 0x00000fa7177ca203   r9 = 0x0000000000000000  r10 = 0x00000f790e3d2285  r11 = 0x0000000000000000\nE   r12 = 0x00000fa7177ca204  r13 = 0x00007da8bbe3f590  r14 = 0x00007da8bbe3f480  r15 = 0x00007d38bbe51018\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\nE   Thread T51 (ydbd.System) created by T0 here:\nE   #0 0x0000211121e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149e555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149e555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf50ec in NActors::TBasicExecutorPool...","1773676288000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 5 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773675081000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773674966000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773674896000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773672978000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 5 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773667338000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 5 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773663833000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773651160000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773644268000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773615765000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773572070000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773562367000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773562210000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773547892000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773510184000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773489674000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773488352000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773485307000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773418209000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773417296000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 5 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773417175000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773413019000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773409896000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 3 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773407266000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 1 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773407147000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773406369000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773405522000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 1 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773404601000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 3 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773402640000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 3 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773401100000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773398367000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 2 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773397572000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 5 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773397083000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773394916000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773394033000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773392201000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773391749000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 3 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773391647000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773389368000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773389248000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773389101000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773388600000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773376892000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 3 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773374483000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773366191000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 5 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773365943000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773364460000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 4 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","1773361816000000":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 5 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count"},"ydb/tests/stress/kv/tests/test_workload.py.TestYdbKvWorkload.test[column]":{"1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:621: in __run\n    self.client.bind_storage_pools(self.domain_name, pools, token=root_token)\nydb/tests/library/clients/kikimr_client.py:197: in bind_storage_pools\n    self.invoke(\nydb/tests/library/clients/kikimr_client.py:122: in invoke\n    return callee(request, timeout=self.__timeout)\ncontrib/python/grpcio/py3/grpc/_channel.py:1181: in __call__\n    return _end_unary_response_blocking(state, call, False, None)\ncontrib/python/grpcio/py3/grpc/_channel.py:1006: in _end_unary_response_blocking\n    raise _InactiveRpcError(state)  # pytype: disable=not-instantiable\nE   grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:\nE   \tstatus = StatusCode.UNAVAILABLE\nE   \tdetails = \"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:2830: Failed to connect to remote host: Connection refused\"\nE   \tdebug_error_string = \"UNKNOWN:Error received from peer  {grpc_message:\"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:2830: Failed to connect to remote host: Connection refused\", grpc_status:14, created_time:\"2026-03-16T16:19:55.375333629+00:00\"}\"\nE   >\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/stress/kv/tests/test_workload.py:12: in setup\n    yield from self.setup_cluster()\nydb/tests/library/stress/fixtures.py:29: in setup_cluster\n    self.cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/kv/tests/test-results/py3test/testing_out_stuff/test_workload.py.TestYdbKvWorkload.test.column/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/kv/tests/test-results/py3test/testing_out_stuff/test_workload.py.TestYdbKvWorkload.test.column/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==1023709==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7b6d97238f40 sp 0x7b6d97238e70 T51)\nE   ==1023709==The signal is caused by a WRITE memory access.\nE   ==1023709==Hint: address points to the zero page.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   AddressSanitizer:DEADLYSIGNAL\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7f6ddf969ac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7f6ddf9fb8bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==1023709==Register values:\nE   rax = 0x00007cddde7f7890  rbx = 0x00007cddde7f78a0  rcx = 0x0000000000000000  rdx = 0x00000f9bbbcfef12\nE   rdi = 0x00007d4dde8316a8  rsi = 0x00000fa9bbd062d5  rbp = 0x00007b6d97238f40  rsp = 0x00007b6d97238e70\nE   r8 = 0x00000f9bbbcfef13   r9 = 0x0000000000000000  r10 = 0x00000f6db2a8e985  r11 = 0x0000000000000000\nE   r12 = 0x00000f9bbbcfef14  r13 = 0x00007d4dde831190  r14 = 0x00007d4dde831080  r15 = 0x00007cddde7f7898\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\nE   Thread T51 (ydbd.System) created by T0 here:\nE   #0 0x0000211121e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149e555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149e555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x000023cf50ec in NActors::T...","1773519197000000":"ydb/tests/stress/kv/tests/test_workload.py:16: in test\n    yatest.common.execute([\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/kv/workload_kv --endpoint grpc://localhost:5106 --database /Root --duration 60 --store_type column --kv_prefix column' has failed with code 1.\nE   Errors:\nE   ...atabase=/Root', 'workload', 'kv', 'run', 'mixed', '--seconds', '60', '--threads', '10', '--cols', '5', '--len', '200', '--int-cols', '2', '--key-cols', '3', '--path', 'column/kv_column']' died with <Signals.SIGABRT: 6>.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/kv/workload/__init__.py\", line 69, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=[\"run\", \"mixed\",\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<5 lines>...\nE                                            \"--key-cols\", \"3\"])\nE                                            ^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/kv/workload/__init__.py\", line 53, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/kv/tests/test-results/py3test/tmpji67b622/kv_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:5106', '--database=/Root', 'workload', 'kv', 'run', 'mixed', '--seconds', '60', '--threads', '10', '--cols', '5', '--len', '200', '--int-cols', '2', '--key-cols', '3', '--path', 'column/kv_column']' died with <Signals.SIGABRT: 6>.","1773445080000000":"Killed by timeout (600 s)"},"ydb/tests/stress/topic/tests/test_workload_topic.py.TestYdbTopicWorkload.test":{"1773729526000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:15839 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...e', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpz09avx9t/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:15839', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   No authentication methods were found; going without authentication","1773722559000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:18446 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpcvq2svo2/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:18446', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773719674000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:25515 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp5jq0m8b4/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:25515', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773717082000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:61992 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpqeke4dp1/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:61992', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773702149000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:61314 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpfvdet5cg/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:61314', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773698398000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:1529 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...n-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpb7mso892/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:1529', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773694511000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:11084 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpkxcu9z3q/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:11084', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773692039000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:5570 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpzdpwrmm1/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:5570', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773690441000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:29462 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ProducerId 98fc100e-216f11f1-b67adc85-e3e316fd PartitionId 50\nE   2026-03-16T19:39:13.784361Z :WARNING: No initial sequence number for ProducerId 994f04b2-216f11f1-8aad3abf-5db46bd2 PartitionId 54\nE   2026-03-16T19:39:13.813288Z :WARNING: No initial sequence number for ProducerId 98d55720-216f11f1-923dfac8-9189d48c PartitionId 51\nE   2026-03-16T19:39:13.813374Z :WARNING: No initial sequence number for ProducerId 992d4cd2-216f11f1-8ab5d654-53a0fba9 PartitionId 48\nE   2026-03-16T19:39:13.815942Z :WARNING: No initial sequence number for ProducerId 99dd5cc6-216f11f1-ae91b8c4-5dfcb255 PartitionId 5\nE   2026-03-16T19:39:13.815988Z :WARNING: No initial sequence number for ProducerId 99feac5a-216f11f1-834ee216-3d757295 PartitionId 8\nE   2026-03-16T19:39:13.816020Z :WARNING: No initial sequence number for ProducerId 99dee6e0-216f11f1-97f3dbff-ffb69fd PartitionId 8\nE   2026-03-16T19:39:13.816080Z :WARNING: No initial sequence number for ProducerId 99ea3bb2-216f11f1-a8a80e88-41f890fc PartitionId 6\nE   2026-03-16T19:39:13.816160Z :WARNING: No initial sequence number for ProducerId 99f15988-216f11f1-912117de-813182cd PartitionId 5\nE   2026-03-16T19:39:13.816231Z :WARNING: No initial sequence number for ProducerId 98b7de66-216f11f1-a45cec97-479645f2 PartitionId 41\nE   2026-03-16T19:39:13.816282Z :WARNING: No initial sequence number for ProducerId 98b9a5a2-216f11f1-944c42a5-d7a074c4 PartitionId 41\nE   2026-03-16T19:39:13.816322Z :WARNING: No initial sequence number for ProducerId 98c5431c-216f11f1-83d75ef5-7dcf4eb4 PartitionId 40","1773690098000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13705 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpcjbs1fb8/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:13705', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773689559000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:2599 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpm3u9gsh8/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:2599', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773689025000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:18386 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpncd6me45/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:18386', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773688476000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:64499 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpyykeznwh/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:64499', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773687448000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:10533 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...y\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmptvx2o5qd/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:10533', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   No authentication methods were found; going without authentication\nE   No authentication methods were found; going without authentication","1773687172000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:9769 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...rgs(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmprrbyphkz/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:9769', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.\nE   2026-03-16T18:47:37.909025Z :WARNING: No initial sequence number for ProducerId 64ca8e52-216811f1-a2366a5d-ebc8f877 PartitionId 1\nE   2026-03-16T18:47:37.913183Z :WARNING: No initial sequence number for ProducerId 64680f84-216811f1-97849edd-b3d009a2 PartitionId 32\nE   2026-03-16T18:47:37.947468Z :WARNING: No initial sequence number for ProducerId 64728f18-216811f1-aedcb5fc-39e1bffc PartitionId 30\nE   2026-03-16T18:47:37.963619Z :WARNING: No initial sequence number for ProducerId 6435883e-216811f1-a066e5e7-1b30b895 PartitionId 25","1773687167000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:10905 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpggyjoevu/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:10905', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773687158000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:24057 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp8gb0_pdz/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:24057', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773687146000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:2341 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...e-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-16T18:45:36.597317Z :WARNING: No initial sequence number for ProducerId 1adc387c-216811f1-ba943eec-73e8cdde PartitionId 25\nE   2026-03-16T18:45:36.616767Z :WARNING: No initial sequence number for ProducerId 1b95c440-216811f1-a3bf62db-5177e5b3 PartitionId 32\nE   2026-03-16T18:45:36.627860Z :WARNING: No initial sequence number for ProducerId 1bb52308-216811f1-a088969a-2f02e8c1 PartitionId 5\nE   2026-03-16T18:45:36.629122Z :WARNING: No initial sequence number for ProducerId 1bb97c96-216811f1-ab925deb-3c28d0c PartitionId 10\nE   2026-03-16T18:45:36.629578Z :WARNING: No initial sequence number for ProducerId 1b7dd952-216811f1-a916010f-ed95ba32 PartitionId 28\nE   2026-03-16T18:45:36.669000Z :WARNING: No initial sequence number for ProducerId 1ae7999c-216811f1-bfb026fd-d3aa67a2 PartitionId 36\nE   2026-03-16T18:45:36.676061Z :WARNING: No initial sequence number for ProducerId 1af905a6-216811f1-9ed0c65c-95480792 PartitionId 29\nE   2026-03-16T18:45:36.687360Z :WARNING: No initial sequence number for ProducerId 1ad3ee10-216811f1-a9a516c3-9f434906 PartitionId 25\nE   2026-03-16T18:45:36.745290Z :WARNING: No initial sequence number for ProducerId 1ad29786-216811f1-918b6c6b-797b5643 PartitionId 28\nE   2026-03-16T18:45:36.767405Z :WARNING: No initial sequence number for ProducerId 1ad1637a-216811f1-94b23840-1f771c79 PartitionId 31","1773685622000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:29856 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...oducerId 4bfd246a-216411f1-b1a30a55-3bbb1d91 PartitionId 2\nE   2026-03-16T18:18:08.857471Z :WARNING: No initial sequence number for ProducerId 4bd6ef7a-216411f1-8de2d253-d399af9 PartitionId 27\nE   2026-03-16T18:18:08.857558Z :WARNING: No initial sequence number for ProducerId 4c065b3e-216411f1-8cae944c-ab49a1e2 PartitionId 28\nE   2026-03-16T18:18:08.857608Z :WARNING: No initial sequence number for ProducerId 4bd17298-216411f1-89e574d7-11dfdd62 PartitionId 25\nE   2026-03-16T18:18:08.857961Z :WARNING: No initial sequence number for ProducerId 4be0f628-216411f1-a16c2856-b77df9d0 PartitionId 28\nE   2026-03-16T18:18:08.858063Z :WARNING: No initial sequence number for ProducerId 4bcf105c-216411f1-b806c609-5d8036b PartitionId 27\nE   2026-03-16T18:18:08.858107Z :WARNING: No initial sequence number for ProducerId 4bc5d884-216411f1-b0bc1982-33d6ce2c PartitionId 27\nE   2026-03-16T18:18:08.870612Z :WARNING: No initial sequence number for ProducerId 4bcd2c24-216411f1-979bb522-43f44c09 PartitionId 0\nE   2026-03-16T18:18:08.873730Z :WARNING: No initial sequence number for ProducerId 4bd5b308-216411f1-8b173ea9-ddf1b0f6 PartitionId 27\nE   2026-03-16T18:18:08.873872Z :WARNING: No initial sequence number for ProducerId 4b28579e-216411f1-8e632ffe-3df5a125 PartitionId 95\nE   2026-03-16T18:18:08.878987Z :WARNING: No initial sequence number for ProducerId 4b0ecc70-216411f1-b02bdc0e-379ade55 PartitionId 93\nE   2026-03-16T18:18:08.889732Z :WARNING: No initial sequence number for ProducerId 4b25a652-216411f1-81997d23-d3730b43 PartitionId 88","1773685521000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13636 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...cerId eab6ad52-216311f1-9a8037c9-41d82bea PartitionId 13\nE   2026-03-16T18:15:20.794494Z :WARNING: No initial sequence number for ProducerId eabb8ce6-216311f1-bb04df46-67c9c641 PartitionId 21\nE   2026-03-16T18:15:20.794544Z :WARNING: No initial sequence number for ProducerId eabfdd78-216311f1-87bbae55-1126537f PartitionId 16\nE   2026-03-16T18:15:20.798798Z :WARNING: No initial sequence number for ProducerId eab83096-216311f1-92c4cbac-f582e9c PartitionId 12\nE   2026-03-16T18:15:20.800933Z :WARNING: No initial sequence number for ProducerId eabcb4cc-216311f1-9676a48a-b5765698 PartitionId 10\nE   2026-03-16T18:15:20.801925Z :WARNING: No initial sequence number for ProducerId eab8605c-216311f1-985cbd15-5d133b16 PartitionId 19\nE   2026-03-16T18:15:20.814763Z :WARNING: No initial sequence number for ProducerId eabf612c-216311f1-928c5419-f5fded3c PartitionId 16\nE   2026-03-16T18:15:20.816724Z :WARNING: No initial sequence number for ProducerId eabb005a-216311f1-a5962117-65b0d9c5 PartitionId 10\nE   2026-03-16T18:15:20.816772Z :WARNING: No initial sequence number for ProducerId eac54592-216311f1-ae698c7e-d92d3705 PartitionId 14\nE   2026-03-16T18:15:20.825746Z :WARNING: No initial sequence number for ProducerId eac417c6-216311f1-bcff5d14-c7193688 PartitionId 11\nE   2026-03-16T18:15:20.826933Z :WARNING: No initial sequence number for ProducerId eb3ba87c-216311f1-b3309538-df75b3f7 PartitionId 21\nE   2026-03-16T18:15:20.831160Z :WARNING: No initial sequence number for ProducerId eb58f878-216311f1-875cd684-a90ad6b1 PartitionId 21","1773683098000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:22202 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpbn3p5f18/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:22202', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-16T17:39:49.707638Z :EMERG: No messages were written.\nE   2026-03-16T17:39:49.759412Z :WARNING: No initial sequence number for ProducerId f0cc235c-215e11f1-b959f0ce-21d91d74 PartitionId 14\nE   2026-03-16T17:39:49.815932Z :WARNING: No initial sequence number for ProducerId f026ee78-215e11f1-a6c7a373-7b095382 PartitionId 99","1773682225000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:3870 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...er for ProducerId 7cd085c-215d11f1-9ef06ad6-95fca6d9 PartitionId 62\nE   2026-03-16T17:26:11.568720Z :WARNING: No initial sequence number for ProducerId 7d0d522-215d11f1-96453c08-b9bd9fee PartitionId 56\nE   2026-03-16T17:26:11.576280Z :WARNING: No initial sequence number for ProducerId 7d1038a-215d11f1-83180a34-ffffed1 PartitionId 48\nE   2026-03-16T17:26:11.589594Z :WARNING: No initial sequence number for ProducerId 7d16960-215d11f1-8a1d988c-5957a293 PartitionId 57\nE   2026-03-16T17:26:11.590009Z :WARNING: No initial sequence number for ProducerId 7d96354-215d11f1-a55ca82d-2d724b8a PartitionId 52\nE   2026-03-16T17:26:11.591434Z :WARNING: No initial sequence number for ProducerId 7cdc56c-215d11f1-91c8ee15-397609fa PartitionId 62\nE   2026-03-16T17:26:11.591493Z :WARNING: No initial sequence number for ProducerId 8019216-215d11f1-8b01a8da-1b6c1046 PartitionId 60\nE   2026-03-16T17:26:11.604730Z :WARNING: No initial sequence number for ProducerId 7e303e6-215d11f1-abc74270-436d44bf PartitionId 47\nE   2026-03-16T17:26:11.604766Z :WARNING: No initial sequence number for ProducerId 7eab938-215d11f1-a04c3edc-29268ba7 PartitionId 63\nE   2026-03-16T17:26:11.604802Z :WARNING: No initial sequence number for ProducerId 7d7b090-215d11f1-88cbbb57-db9dd502 PartitionId 57\nE   2026-03-16T17:26:11.608604Z :WARNING: No initial sequence number for ProducerId 7d15646-215d11f1-aab377ff-f31dfe5a PartitionId 45\nE   2026-03-16T17:26:11.609593Z :WARNING: No initial sequence number for ProducerId 7eabda2-215d11f1-a34f708b-d5dc3710 PartitionId 64","1773682040000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13830 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...   self.run_topic_write_with_tx(20, 5, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmptj0_wf8l/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:13830', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p5_pq10']' returned non-zero exit status 1.\nE   2026-03-16T17:19:05.861253Z :WARNING: No initial sequence number for ProducerId 2e18af8-215c11f1-ae8062de-8519b3ce PartitionId 61\nE   2026-03-16T17:19:05.865039Z :WARNING: No initial sequence number for ProducerId 35b192c-215c11f1-9e7f9cc4-6f89402a PartitionId 57","1773681818000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:28637 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmprltzxkj9/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:28637', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773680250000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:9997 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...te-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp7tqq651i/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:9997', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   2026-03-16T15:31:09.315809Z :EMERG: No messages were written.","1773679933000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23575 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpxugc1pdd/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:23575', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773679449000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:17556 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp31tob2ff/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:17556', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773679120000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:27530 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp6bkp9c_f/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:27530', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773679007000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:19852 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...rocessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmphf6twon1/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:19852', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-16T16:30:38.435067Z :WARNING: No initial sequence number for ProducerId 42a8c496-215511f1-9e9aed58-e387bf35 PartitionId 60\nE   2026-03-16T16:30:38.446823Z :WARNING: No initial sequence number for ProducerId 42ac1e7a-215511f1-849268bd-bb37fa9e PartitionId 53\nE   2026-03-16T16:30:38.502277Z :WARNING: No initial sequence number for ProducerId 43d7d2a8-215511f1-a9511803-47b7fb6f PartitionId 16\nE   2026-03-16T16:30:38.541567Z :WARNING: No initial sequence number for ProducerId 43e1f45e-215511f1-837b6a6c-f11bf12 PartitionId 17\nE   2026-03-16T16:30:38.541725Z :WARNING: No initial sequence number for ProducerId 43e6c240-215511f1-84b8d831-19c2ba29 PartitionId 21\nE   2026-03-16T16:30:38.580579Z :WARNING: No initial sequence number for ProducerId 4331546e-215511f1-ae1ca2ab-17167319 PartitionId 26\nE   2026-03-16T16:30:38.583621Z :WARNING: No initial sequence number for ProducerId 43f27234-215511f1-9d7acb55-c55f4006 PartitionId 21","1773678560000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:12915 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmps6u7rsoz/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:12915', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773678541000000":"setup failed:\nydb/tests/library/harness/kikimr_runner.py:540: in start\n    self.__run(timeout_seconds=timeout_seconds)\nydb/tests/library/harness/kikimr_runner.py:621: in __run\n    self.client.bind_storage_pools(self.domain_name, pools, token=root_token)\nydb/tests/library/clients/kikimr_client.py:197: in bind_storage_pools\n    self.invoke(\nydb/tests/library/clients/kikimr_client.py:122: in invoke\n    return callee(request, timeout=self.__timeout)\ncontrib/python/grpcio/py3/grpc/_channel.py:1181: in __call__\n    return _end_unary_response_blocking(state, call, False, None)\ncontrib/python/grpcio/py3/grpc/_channel.py:1006: in _end_unary_response_blocking\n    raise _InactiveRpcError(state)  # pytype: disable=not-instantiable\nE   grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:\nE   \tstatus = StatusCode.UNAVAILABLE\nE   \tdetails = \"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:22769: Failed to connect to remote host: Connection refused\"\nE   \tdebug_error_string = \"UNKNOWN:Error received from peer  {created_time:\"2026-03-16T16:19:05.508859114+00:00\", grpc_status:14, grpc_message:\"failed to connect to all addresses; last error: UNKNOWN: ipv4:127.0.0.1:22769: Failed to connect to remote host: Connection refused\"}\"\nE   >\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/stress/topic/tests/test_workload_topic.py:12: in setup\n    yield from self.setup_cluster()\nydb/tests/library/stress/fixtures.py:29: in setup_cluster\n    self.cluster.start()\nydb/tests/library/harness/kikimr_runner.py:545: in start\n    self.stop()\nydb/tests/library/harness/kikimr_runner.py:766: in stop\n    raise daemon.SeveralDaemonErrors(saved_exceptions)\nE   ydb.tests.library.harness.daemon.SeveralDaemonErrors: Daemon failed with message: Unexpectedly finished before stop.\nE   Process exit_code = 100.\nE   Stdout file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/testing_out_stuff/test_workload_topic.py.TestYdbTopicWorkload.test/cluster/node_1/stdout\nE   Stderr file name: \nE   /home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/testing_out_stuff/test_workload_topic.py.TestYdbTopicWorkload.test/cluster/node_1/stderr\nE   Stderr content:\nE   \nE   GRpc memory quota was set but disabled due to issues with grpc quoter, to enable it use EnableGRpcMemoryQuota option\nE   AddressSanitizer:DEADLYSIGNAL\nE   =================================================================\nE   ==1020895==ERROR: AddressSanitizer: SEGV on unknown address 0x000000000018 (pc 0x000042edf29e bp 0x7b987750af40 sp 0x7b987750ae70 T52)\nE   ==1020895==The signal is caused by a WRITE memory access.\nE   ==1020895==Hint: address points to the zero page.\nE   warning: address range table at offset 0x10c0 has a premature terminator entry at offset 0x10d0\nE   AddressSanitizer:DEADLYSIGNAL\nE   #0 0x000042edf29e in AtomicAdd /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12\nE   #1 0x000042edf29e in Add /-S/library/cpp/monlib/counters/counters.h:78:20\nE   #2 0x000042edf29e in operator+= /-S/library/cpp/monlib/counters/counters.h:100:13\nE   #3 0x000042edf29e in Delta /-S/ydb/core/blobstorage/vdisk/common/memusage.h:36:22\nE   #4 0x000042edf29e in Subtract /-S/ydb/core/blobstorage/vdisk/common/memusage.h:32:13\nE   #5 0x000042edf29e in ~TMemoryConsumerWithDropOnDestroy /-S/ydb/core/blobstorage/vdisk/common/memusage.h:56:30\nE   #6 0x000042edf29e in reset /-S/contrib/libs/cxxsupp/libcxx/include/optional:319:15\nE   #7 0x000042edf29e in TUnreplicatedBlobRecords /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.h:186:32\nE   #8 0x000042edf29e in NKikimr::THullReplJobActor::THullReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>&&, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1083:15\nE   #9 0x000042ede8e7 in NKikimr::CreateReplJobActor(std::__y1::shared_ptr<NKikimr::TReplCtx>, NActors::TActorId const&, NKikimr::TLogoBlobID const&, std::__y1::shared_ptr<THashMap<NKikimr::TVDiskIdShort, NActors::TActorId, THash<NKikimr::TVDiskIdShort>, TEqualTo<NKikimr::TVDiskIdShort>, std::__y1::allocator<NKikimr::TVDiskIdShort>>>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::shared_ptr<NKikimr::TBlobIdQueue>, std::__y1::optional<std::__y1::pair<NKikimr::TVDiskID, NActors::TActorId>> const&, NKikimr::TUnreplicatedBlobRecords&&, NKikimr::TMilestoneQueue&&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_hullrepljob.cpp:1120:20\nE   #10 0x000042ecd0a9 in NKikimr::TReplScheduler::RunRepl(NKikimr::TLogoBlobID const&) /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:536:39\nE   #11 0x000042eb832e in NKikimr::TReplScheduler::StartReplication() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:304:13\nE   #12 0x000042eb59a6 in NKikimr::TReplScheduler::Bootstrap() /-S/ydb/core/blobstorage/vdisk/repl/blobstorage_repl.cpp:260:17\nE   #13 0x000042eb38ce in NActors::TActorBootstrapped<NKikimr::TReplScheduler>::StateBootstrap(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor_bootstrapped.h:26:22\nE   #14 0x000023c342a7 in NActors::IActor::Receive(TAutoPtr<NActors::IEventHandle, TDelete>&) /-S/ydb/library/actors/core/actor.cpp:354:17\nE   #15 0x000023d261b1 in NActors::TExecutorThread::Execute(NActors::TMailbox*, bool) /-S/ydb/library/actors/core/executor_thread.cpp:267:28\nE   #16 0x000023d2fdc6 in NActors::TExecutorThread::ProcessExecutorPool()::$_0::operator()(NActors::TMailbox*, bool) const /-S/ydb/library/actors/core/executor_thread.cpp:455:39\nE   #17 0x000023d2f37d in NActors::TExecutorThread::ProcessExecutorPool() /-S/ydb/library/actors/core/executor_thread.cpp:507:13\nE   #18 0x000023d313ee in NActors::TExecutorThread::ThreadProc() /-S/ydb/library/actors/core/executor_thread.cpp:533:9\nE   #19 0x0000214adb54 in (anonymous namespace)::TPosixThread::ThreadProxy(void*) /-S/util/system/thread.cpp:245:20\nE   #20 0x00002112b5a6 in asan_thread_start(void*) /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:239:28\nE   #21 0x7f98c046fac2  (/lib/x86_64-linux-gnu/libc.so.6+0x94ac2) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   #22 0x7f98c05018bf  (/lib/x86_64-linux-gnu/libc.so.6+0x1268bf) (BuildId: 4f7b0c955c3d81d7cac1501a2498b69d1d82bfe7)\nE   \nE   ==1020895==Register values:\nE   rax = 0x00007d08bf2f3990  rbx = 0x00007d08bf2f39a0  rcx = 0x0000000000000000  rdx = 0x00000fa117e5e732\nE   rdi = 0x00007d78bf3316a8  rsi = 0x00000faf17e662d5  rbp = 0x00007b987750af40  rsp = 0x00007b987750ae70\nE   r8 = 0x00000fa117e5e733   r9 = 0x0000000000000000  r10 = 0x00000f730e988985  r11 = 0x0000000000000000\nE   r12 = 0x00000fa117e5e734  r13 = 0x00007d78bf331190  r14 = 0x00007d78bf331080  r15 = 0x00007d08bf2f3998\nE   AddressSanitizer can not provide additional info.\nE   SUMMARY: AddressSanitizer: SEGV /-S/library/cpp/deprecated/atomic/atomic_gcc.h:51:12 in AtomicAdd\nE   Thread T52 (ydbd.System) created by T0 here:\nE   #0 0x0000211121e1 in pthread_create /-S/contrib/libs/clang20-rt/lib/asan/asan_interceptors.cpp:250:3\nE   #1 0x00002149e555 in Start /-S/util/system/thread.cpp:230:27\nE   #2 0x00002149e555 in TThread::Start() /-S/util/system/thread.cpp:315:34\nE   #3 0x00002...","1773678224000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:18300 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...n.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpo0v0wy4e/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:18300', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-16T16:18:57.249591Z :WARNING: No initial sequence number for ProducerId 9f169930-215311f1-88781633-cb574650 PartitionId 66","1773677952000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:21071 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpoz7nip21/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:21071', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773677718000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:22391 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmptu5hsi3_/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:22391', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-16T16:06:25.624547Z :WARNING: No initial sequence number for ProducerId d9974fa2-215111f1-95e50248-e533b0b6 PartitionId 54\nE   2026-03-16T16:06:25.630522Z :WARNING: No initial sequence number for ProducerId d9965e3a-215111f1-9cf39390-39b35de PartitionId 48\nE   2026-03-16T16:06:25.650785Z :WARNING: No initial sequence number for ProducerId d9a7c08a-215111f1-a4398f47-f9a53e53 PartitionId 51\nE   2026-03-16T16:06:25.665482Z :WARNING: No initial sequence number for ProducerId d9a632ba-215111f1-a43ac299-9ffda0fe PartitionId 51\nE   2026-03-16T16:06:25.801247Z :WARNING: No initial sequence number for ProducerId d9962c44-215111f1-9d4e9263-f321fc80 PartitionId 48","1773677510000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:3507 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp0ng_v1oi/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:3507', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773677230000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:5408 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpg89j2ovr/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:5408', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773677073000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:27250 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpjmm4fsqj/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:27250', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773676288000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:9479 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...kload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpd9gy0ttb/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:9479', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773675427000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:9374 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...kload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp0gsak28b/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:9374', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773675081000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:25545 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpxhzljkuv/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:25545', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773674966000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:18970 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...cerId 7cea226c-214b11f1-a3030348-adec7875 PartitionId 20\nE   2026-03-16T15:20:58.240878Z :WARNING: No initial sequence number for ProducerId 7d027bbe-214b11f1-b16c9f3c-d55bd26b PartitionId 18\nE   2026-03-16T15:20:58.251519Z :WARNING: No initial sequence number for ProducerId 7bcca350-214b11f1-91c73bae-f9f2c9e7 PartitionId 33\nE   2026-03-16T15:20:58.254475Z :WARNING: No initial sequence number for ProducerId 7bda9afa-214b11f1-b26a6dc6-5da53714 PartitionId 28\nE   2026-03-16T15:20:58.254547Z :WARNING: No initial sequence number for ProducerId 7b7fa6e0-214b11f1-b04fb7f9-2542ab08 PartitionId 23\nE   2026-03-16T15:20:58.292462Z :WARNING: No initial sequence number for ProducerId 7c119d20-214b11f1-b379bcb8-6bd6798d PartitionId 69\nE   2026-03-16T15:20:58.299468Z :WARNING: No initial sequence number for ProducerId 7c33b48c-214b11f1-b93c2d5d-3be0c2c5 PartitionId 33\nE   2026-03-16T15:20:58.304135Z :WARNING: No initial sequence number for ProducerId 7c33c15c-214b11f1-88790900-470bf8f0 PartitionId 38\nE   2026-03-16T15:20:58.307782Z :WARNING: No initial sequence number for ProducerId 7b8560b2-214b11f1-a3f88226-3bc73885 PartitionId 72\nE   2026-03-16T15:20:58.312905Z :WARNING: No initial sequence number for ProducerId 7c16d182-214b11f1-82231339-7495d43 PartitionId 35\nE   2026-03-16T15:20:58.313021Z :WARNING: No initial sequence number for ProducerId 7bf78b24-214b11f1-99c5cb16-3b81a2e1 PartitionId 71\nE   2026-03-16T15:20:58.331625Z :WARNING: No initial sequence number for ProducerId 7bea0800-214b11f1-9b11002f-451bd2da PartitionId 29","1773674896000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:30176 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpf1x9w5gm/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:30176', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-16T15:18:57.931609Z :WARNING: No initial sequence number for ProducerId 38ed3874-214b11f1-99521d72-fb138f44 PartitionId 8\nE   2026-03-16T15:18:57.931681Z :WARNING: No initial sequence number for ProducerId 38c1ccfc-214b11f1-bdb80ccd-438f989a PartitionId 21\nE   2026-03-16T15:18:57.939453Z :WARNING: No initial sequence number for ProducerId 38c6428c-214b11f1-82dc48f2-9164d719 PartitionId 17\nE   2026-03-16T15:18:58.088549Z :WARNING: No initial sequence number for ProducerId 38c61a3c-214b11f1-a1bc1a37-6736a341 PartitionId 17\nE   2026-03-16T15:18:58.134997Z :WARNING: No initial sequence number for ProducerId 39fd50aa-214b11f1-a4258677-337e51b8 PartitionId 11","1773674709000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:8220 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp3x127qyg/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:8220', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773674432000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:27189 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpcririjug/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:27189', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773672978000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:4988 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...kload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpglbmopdd/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:4988', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773672747000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:11945 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...3', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpsvjdhpf1/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:11945', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   2026-03-16T14:45:37.855150Z :WARNING: No initial sequence number for ProducerId 96e63674-214611f1-b70730c6-dda581bf PartitionId 34","1773671187000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:21123 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...   ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp1j4q92cs/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:21123', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.\nE   2026-03-16T14:18:03.321185Z :WARNING: No initial sequence number for ProducerId c184e212-214211f1-a2c29ddb-97eaff86 PartitionId 7","1773668553000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:64157 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpmrhx19vk/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:64157', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773668400000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:10154 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmponwqj_ny/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:10154', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773667338000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:28863 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmphh5kbg40/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:28863', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773667038000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23201 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpu8ftxf48/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:23201', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773666791000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:11195 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ucerId c645f46c-213811f1-b0b37ea6-abaea495 PartitionId 32\nE   2026-03-16T13:06:26.953708Z :WARNING: No initial sequence number for ProducerId c6456542-213811f1-ac56ea03-c7666d76 PartitionId 0\nE   2026-03-16T13:06:26.980371Z :WARNING: No initial sequence number for ProducerId c5084438-213811f1-8fd677e4-145190d PartitionId 13\nE   2026-03-16T13:06:27.004347Z :WARNING: No initial sequence number for ProducerId c65738da-213811f1-8f2d92fb-e312453d PartitionId 29\nE   2026-03-16T13:06:27.008552Z :WARNING: No initial sequence number for ProducerId c5067afe-213811f1-886db45c-17b6d03b PartitionId 17\nE   2026-03-16T13:06:27.018216Z :WARNING: No initial sequence number for ProducerId c5683c08-213811f1-ab197c5d-21a4f558 PartitionId 75\nE   2026-03-16T13:06:27.022095Z :WARNING: No initial sequence number for ProducerId c523b7d6-213811f1-8cbe640a-6109272b PartitionId 21\nE   2026-03-16T13:06:27.031802Z :WARNING: No initial sequence number for ProducerId c572b7c8-213811f1-91ce1d71-35d79b37 PartitionId 78\nE   2026-03-16T13:06:27.035850Z :WARNING: No initial sequence number for ProducerId c572fa6c-213811f1-aed911e0-bb4efd76 PartitionId 72\nE   2026-03-16T13:06:27.051895Z :WARNING: No initial sequence number for ProducerId c6495b8e-213811f1-af014d23-d19d097e PartitionId 30\nE   2026-03-16T13:06:27.052515Z :WARNING: No initial sequence number for ProducerId c64ef9fe-213811f1-82649bd5-3d1ae555 PartitionId 32\nE   2026-03-16T13:06:27.053522Z :WARNING: No initial sequence number for ProducerId c57a41c8-213811f1-a5ec4e39-75a61afc PartitionId 85","1773665446000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23664 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...s/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpfwff3xq7/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:23664', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-16T12:41:20.995586Z :EMERG: No messages were written.","1773663833000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23595 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...te-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.\nE   2026-03-16T12:17:00.831661Z :WARNING: No initial sequence number for ProducerId d15154c0-213111f1-a0c577f4-a706847e PartitionId 17\nE   2026-03-16T12:17:00.831661Z :WARNING: No initial sequence number for ProducerId d14b7e4c-213111f1-ad9f499b-11d1d7b0 PartitionId 14\nE   2026-03-16T12:17:00.836283Z :WARNING: No initial sequence number for ProducerId d11047f0-213111f1-ab1b3b48-7b5bbeaf PartitionId 16\nE   2026-03-16T12:17:00.847248Z :WARNING: No initial sequence number for ProducerId d14a4cde-213111f1-8452d51b-dfb9e561 PartitionId 17\nE   2026-03-16T12:17:00.850872Z :WARNING: No initial sequence number for ProducerId d0f723ce-213111f1-9ff8d2c1-cd7d739e PartitionId 6\nE   2026-03-16T12:17:00.851499Z :WARNING: No initial sequence number for ProducerId d0fe6c88-213111f1-bb05a5c2-21088d89 PartitionId 10\nE   2026-03-16T12:17:00.857024Z :WARNING: No initial sequence number for ProducerId d1285c50-213111f1-b364e919-8b8ae401 PartitionId 20\nE   2026-03-16T12:17:00.860358Z :WARNING: No initial sequence number for ProducerId d109eebe-213111f1-99d64a01-2f52019b PartitionId 17\nE   2026-03-16T12:17:00.862600Z :WARNING: No initial sequence number for ProducerId d0f751f0-213111f1-9a94c1d2-551b3e88 PartitionId 7\nE   2026-03-16T12:17:00.866596Z :WARNING: No initial sequence number for ProducerId d11087a6-213111f1-82054020-89b71e31 PartitionId 18","1773661068000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13190 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...10']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 71, in __two_tablets_distributed_transaction\nE       self.run_topic_write_with_tx(20, 5, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmps3qljeti/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:13190', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p5_pq10']' returned non-zero exit status 1.","1773660879000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:22295 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...s/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpwdj723sc/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:22295', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-16T11:28:56.126563Z :EMERG: No messages were written.","1773659987000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:5530 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpuadxamzb/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:5530', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773659879000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:26728 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpr4uz1xkv/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:26728', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773655409000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23640 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpn66ok0tl/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:23640', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773653768000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:21962 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp7c0kd01h/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:21962', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773653670000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13611 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpnpdkk1nz/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:13611', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773651160000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:28233 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp_u38omzj/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:28233', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p5_pq10']' returned non-zero exit status 1.\nE   2026-03-16T08:46:28.252651Z :WARNING: No initial sequence number for ProducerId 705cf5ce-211411f1-8c3622e7-f4a4ef1 PartitionId 74\nE   2026-03-16T08:46:28.348659Z :WARNING: No initial sequence number for ProducerId 7041f724-211411f1-89e1fe07-2772ba0e PartitionId 77\nE   2026-03-16T08:46:28.371030Z :WARNING: No initial sequence number for ProducerId 705f5634-211411f1-9b42b0b6-36be6cf PartitionId 33","1773647899000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:29644 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...^^^^^^^^^^^^^^^\nE       ))\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 264, in run_topic_write_with_tx\nE       self._run_workload(\nE       ~~~~~~~~~~~~~~~~~~^\nE           topic_name,\nE           ^^^^^^^^^^^\nE       ...<6 lines>...\nE           with_config=True\nE           ^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 140, in _run_workload\nE       self.cmd_run_with_monitoring(self.get_command_prefix(subcmds=args))\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 153, in cmd_run_with_monitoring\nE       executor.run_with_monitoring(cmd)\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^\nE     File \"ydb/tests/stress/topic/workload/command_executor.py\", line 56, in run_with_monitoring\nE       raise subprocess.CalledProcessError(\nE       ...<3 lines>...\nE       )\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpayj8bazg/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:29644', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--window=30', '--byte-rate', '10M', '-p', '33', '-c', '16', '--topic', 'workload_topic_pr33_p1_pq1', '-t', '16', '--use-tx', '--tx-commit-interval', '2000', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--configure-consumers', '500', '--describe-topic', '--describe-consumer', 'data_holder']' returned non-zero exit status 1.","1773644268000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:16490 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpbwpfkl8u/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:16490', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773643731000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23073 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpm5yelx_h/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:23073', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773642105000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:14774 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ucerId d7420010-20fe11f1-b0fae9ae-3d30b9a9 PartitionId 13\nE   2026-03-16T06:11:55.606993Z :WARNING: No initial sequence number for ProducerId d73c1b78-20fe11f1-b25e5977-6701df57 PartitionId 13\nE   2026-03-16T06:11:55.607827Z :WARNING: No initial sequence number for ProducerId d73a37f4-20fe11f1-8cb5741b-bb278daf PartitionId 13\nE   2026-03-16T06:11:55.609300Z :WARNING: No initial sequence number for ProducerId d738617c-20fe11f1-a1a03d1f-d941783 PartitionId 13\nE   2026-03-16T06:11:55.611612Z :WARNING: No initial sequence number for ProducerId d7386d20-20fe11f1-93ed3012-5322ba38 PartitionId 13\nE   2026-03-16T06:11:55.620672Z :WARNING: No initial sequence number for ProducerId d74dfd8e-20fe11f1-ae6190ac-b2c907a PartitionId 13\nE   2026-03-16T06:11:55.611674Z :WARNING: No initial sequence number for ProducerId d73aa392-20fe11f1-8bc27c5b-87db96c1 PartitionId 13\nE   2026-03-16T06:11:55.613192Z :WARNING: No initial sequence number for ProducerId d7386d48-20fe11f1-98aec409-cd322b50 PartitionId 13\nE   2026-03-16T06:11:55.613407Z :WARNING: No initial sequence number for ProducerId d73a7570-20fe11f1-8a8c53dc-3541d70b PartitionId 13\nE   2026-03-16T06:11:55.615347Z :WARNING: No initial sequence number for ProducerId d74c6b68-20fe11f1-8432a792-afd31838 PartitionId 13\nE   2026-03-16T06:11:55.638911Z :WARNING: No initial sequence number for ProducerId d7394b64-20fe11f1-aa352e03-3bc6e71d PartitionId 13\nE   2026-03-16T06:11:55.638996Z :WARNING: No initial sequence number for ProducerId d7495464-20fe11f1-8f977f59-37fd13a1 PartitionId 13","1773637210000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:1691 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpmlgmld06/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:1691', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773636520000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:7912 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpl2l4ivlk/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:7912', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773634122000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:24338 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpfdmj2qix/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:24338', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773628100000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:28843 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpycjnzn8u/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:28843', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773627098000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:6225 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp85orekea/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:6225', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773626829000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:24042 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...s/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpdi4pqigl/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:24042', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-16T02:02:58.323182Z :EMERG: No messages were written.","1773624453000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:31222 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp62_ril9v/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:31222', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773622583000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:65343 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpnsgafoa2/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:65343', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773620795000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:12883 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpf0sjy4yz/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:12883', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773617923000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13063 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp3elp66f5/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:13063', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773615765000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:27852 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...le \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp_xhqiq6w/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:27852', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-15T22:51:27.366386Z :WARNING: No initial sequence number for ProducerId 51187eb6-20c111f1-845a91c5-7d713c8b PartitionId 25","1773615308000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:61737 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpkwmmj9h3/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:61737', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773608284000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:29231 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...erId 5b22cd0a-20b011f1-93c7b047-b5de7187 PartitionId 26\nE   2026-03-15T20:50:15.736223Z :WARNING: No initial sequence number for ProducerId 5b0f801a-20b011f1-b2ac70e4-9174dee1 PartitionId 24\nE   2026-03-15T20:50:15.747962Z :WARNING: No initial sequence number for ProducerId 5b21deb8-20b011f1-b67c99e6-a5118dc4 PartitionId 24\nE   2026-03-15T20:50:15.748438Z :WARNING: No initial sequence number for ProducerId 5b1ae6c6-20b011f1-97f8fd1f-b5c7a3f7 PartitionId 24\nE   2026-03-15T20:50:15.748527Z :WARNING: No initial sequence number for ProducerId 5b12e642-20b011f1-b8342f22-7be309fa PartitionId 24\nE   2026-03-15T20:50:15.754995Z :WARNING: No initial sequence number for ProducerId 5b2b6172-20b011f1-972054a6-f7af0a62 PartitionId 26\nE   2026-03-15T20:50:15.756256Z :WARNING: No initial sequence number for ProducerId 5b254062-20b011f1-88666ba6-a54dd2c0 PartitionId 25\nE   2026-03-15T20:50:15.767528Z :WARNING: No initial sequence number for ProducerId 5b2df1f8-20b011f1-99203a5b-953be0f3 PartitionId 30\nE   2026-03-15T20:50:15.785423Z :WARNING: No initial sequence number for ProducerId 5b22c026-20b011f1-8af15b0b-9fb223fe PartitionId 24\nE   2026-03-15T20:50:15.809670Z :WARNING: No initial sequence number for ProducerId 5b1fbfa2-20b011f1-a7d3734f-793b3463 PartitionId 28\nE   2026-03-15T20:50:15.823248Z :WARNING: No initial sequence number for ProducerId 5a14242c-20b011f1-b0e9ee61-29cbe6ba PartitionId 85\nE   2026-03-15T20:50:15.836973Z :WARNING: No initial sequence number for ProducerId 5a193da4-20b011f1-bf07248f-7d016c61 PartitionId 95","1773606561000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:30538 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpcd3a86ix/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:30538', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773592886000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:24059 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmprh2c96sj/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:24059', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773572070000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:16385 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...on.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp2ex5z29t/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:16385', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-15T10:49:24.053019Z :WARNING: No initial sequence number for ProducerId 6de26562-205c11f1-b4d276ec-4143fed5 PartitionId 5","1773570977000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:21227 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...lf.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpa_c8q4d1/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:21227', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.\nE   2026-03-15T10:31:44.204818Z :WARNING: No initial sequence number for ProducerId f21e67b6-205911f1-b9208298-fd35f2b1 PartitionId 14\nE   2026-03-15T10:31:44.263631Z :WARNING: No initial sequence number for ProducerId f27dcaee-205911f1-8bec097f-e3c045ce PartitionId 23","1773562367000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:7839 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...elf.run_topic_write_with_tx(20, 10, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp_u66i2_k/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:7839', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p10_pq10']' returned non-zero exit status 1.\nE   2026-03-15T08:04:16.993579Z :WARNING: No initial sequence number for ProducerId 5892c3d0-204511f1-976da879-91cd35bd PartitionId 23\nE   2026-03-15T08:04:17.031099Z :WARNING: No initial sequence number for ProducerId 58818b42-204511f1-8150bbf8-1df59ba2 PartitionId 21","1773562210000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:1915 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ..._args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp3ju3nezy/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:1915', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-15T07:56:21.047449Z :WARNING: No initial sequence number for ProducerId 409d35e0-204411f1-802e4fa9-f7798044 PartitionId 94\nE   2026-03-15T07:56:21.178888Z :WARNING: No initial sequence number for ProducerId 41901d14-204411f1-9579d935-41d81846 PartitionId 23\nE   2026-03-15T07:56:21.180571Z :WARNING: No initial sequence number for ProducerId 417ce69a-204411f1-9ee0ff66-71fbddaa PartitionId 22\nE   2026-03-15T07:56:21.184648Z :WARNING: No initial sequence number for ProducerId 41381178-204411f1-a10cd1d2-818d07fa PartitionId 12","1773550610000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:18692 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp10v0ldxe/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:18692', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773549981000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:64805 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpefnm9i1x/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:64805', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773547892000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:14213 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpspg8r4bh/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:14213', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773533668000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:26183 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpbklpdt47/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:26183', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773532000000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:10405 --database /Root --duration 180 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...d non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 68, in __one_tablet_but_a_distributed_transaction\nE       self.run_topic_write_with_tx(20, 10, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmprhp1b2yh/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:10405', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '180', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p10_pq10']' returned non-zero exit status 100.","1773531735000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:62095 --database /Root --duration 180 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...d non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 68, in __one_tablet_but_a_distributed_transaction\nE       self.run_topic_write_with_tx(20, 10, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpat3_xngq/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:62095', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '180', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p10_pq10']' returned non-zero exit status 100.","1773531294000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:17378 --database /Root --duration 180 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...us 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp6_ape6fh/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:17378', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '180', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 100.","1773526956000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:7765 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ile \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpv7ryy3z2/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:7765', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-14T22:19:20.080312Z :WARNING: No initial sequence number for ProducerId a20f3c52-1ff311f1-92d63f3c-ed23ce44 PartitionId 22","1773519197000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:3973 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...n-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpxekvdhae/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:3973', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773517997000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:24448 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpizizomnw/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:24448', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773510184000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:16222 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...erId 97e12c86-1fcb11f1-972326b0-9dd34d41 PartitionId 29\nE   2026-03-14T17:32:19.751245Z :WARNING: No initial sequence number for ProducerId 97e60c24-1fcb11f1-a80f5500-c7f4e192 PartitionId 29\nE   2026-03-14T17:32:19.762176Z :WARNING: No initial sequence number for ProducerId 97e4439e-1fcb11f1-9ee43701-d7269ccc PartitionId 29\nE   2026-03-14T17:32:19.762249Z :WARNING: No initial sequence number for ProducerId 981081ac-1fcb11f1-bcbd8e8f-25221938 PartitionId 29\nE   2026-03-14T17:32:19.766116Z :WARNING: No initial sequence number for ProducerId 97e1801e-1fcb11f1-ab76a291-c50f7335 PartitionId 29\nE   2026-03-14T17:32:19.794127Z :WARNING: No initial sequence number for ProducerId 9823f71e-1fcb11f1-9974bf37-cd0b4fe7 PartitionId 29\nE   2026-03-14T17:32:19.798717Z :WARNING: No initial sequence number for ProducerId 97e1b1f6-1fcb11f1-b78d13f2-d904eb8e PartitionId 29\nE   2026-03-14T17:32:19.809981Z :WARNING: No initial sequence number for ProducerId 97e355ec-1fcb11f1-b2146fb3-2fc54de5 PartitionId 29\nE   2026-03-14T17:32:19.821019Z :WARNING: No initial sequence number for ProducerId 97e69964-1fcb11f1-93022b94-a195839d PartitionId 29\nE   2026-03-14T17:32:19.821093Z :WARNING: No initial sequence number for ProducerId 98184856-1fcb11f1-a4e17888-7b772fcd PartitionId 29\nE   2026-03-14T17:32:19.821135Z :WARNING: No initial sequence number for ProducerId 97f184d2-1fcb11f1-a75be303-e90eb0ff PartitionId 29\nE   2026-03-14T17:32:19.821174Z :WARNING: No initial sequence number for ProducerId 97ea8e66-1fcb11f1-8dafcc81-edf4f06c PartitionId 29","1773508856000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:10220 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...b_cli', '--verbose', '--endpoint', 'grpc://localhost:10220', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.\nE   2026-03-14T17:12:29.517070Z :WARNING: No initial sequence number for ProducerId c42e0118-1fc811f1-81af527e-9949adf9 PartitionId 5\nE   2026-03-14T17:12:29.535916Z :WARNING: No initial sequence number for ProducerId c42b75ec-1fc811f1-b503581d-39e3bdaf PartitionId 7\nE   2026-03-14T17:12:29.551843Z :WARNING: No initial sequence number for ProducerId c40ecb22-1fc811f1-9ad8b791-1daf2799 PartitionId 5\nE   2026-03-14T17:12:29.575971Z :WARNING: No initial sequence number for ProducerId c3568576-1fc811f1-a2f26a90-138f80d9 PartitionId 97\nE   2026-03-14T17:12:29.580453Z :WARNING: No initial sequence number for ProducerId c40ba884-1fc811f1-a10c6c4a-d5cc162e PartitionId 7\nE   2026-03-14T17:12:29.586274Z :WARNING: No initial sequence number for ProducerId c365d116-1fc811f1-8b1ebee7-2fed189b PartitionId 88\nE   2026-03-14T17:12:29.604201Z :WARNING: No initial sequence number for ProducerId c431ac00-1fc811f1-b47061fb-43671af4 PartitionId 9\nE   2026-03-14T17:12:29.623720Z :WARNING: No initial sequence number for ProducerId c46edbfc-1fc811f1-be2deb5f-9d9d0ae0 PartitionId 11\nE   2026-03-14T17:12:29.632545Z :WARNING: No initial sequence number for ProducerId c418f1ce-1fc811f1-8c976902-c90035cd PartitionId 3","1773495254000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:27576 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp9xemons8/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:27576', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773495175000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:6510 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...kload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp6fn_wj5m/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:6510', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773494895000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:14658 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp6kr4pc6t/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:14658', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773489674000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:10241 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp1_jsq1tg/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:10241', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773488352000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:28299 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...li', '--verbose', '--endpoint', 'grpc://localhost:28299', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-14T11:27:21.450487Z :WARNING: No initial sequence number for ProducerId 888cfe0a-1f9811f1-94d2a6a6-31e4accf PartitionId 65\nE   2026-03-14T11:27:21.490326Z :WARNING: No initial sequence number for ProducerId 88afd3d0-1f9811f1-913590ef-79cb059d PartitionId 61\nE   2026-03-14T11:27:21.504136Z :WARNING: No initial sequence number for ProducerId 88a3d40e-1f9811f1-8cc1c4a2-f15b3f1f PartitionId 61\nE   2026-03-14T11:27:21.504233Z :WARNING: No initial sequence number for ProducerId 884625a2-1f9811f1-a992ec07-35e478b3 PartitionId 50\nE   2026-03-14T11:27:21.542922Z :WARNING: No initial sequence number for ProducerId 886f97ca-1f9811f1-8ae3797b-2f932b80 PartitionId 55\nE   2026-03-14T11:27:21.542992Z :WARNING: No initial sequence number for ProducerId 8891e53c-1f9811f1-86c3f690-753f4067 PartitionId 53\nE   2026-03-14T11:27:21.543144Z :WARNING: No initial sequence number for ProducerId 8939bd52-1f9811f1-bb52ea19-2935081e PartitionId 15\nE   2026-03-14T11:27:21.577434Z :WARNING: No initial sequence number for ProducerId 8936c886-1f9811f1-9c040788-3f80b104 PartitionId 17\nE   2026-03-14T11:27:21.682049Z :WARNING: No initial sequence number for ProducerId 893cc38a-1f9811f1-b21705b8-eb032e15 PartitionId 15","1773488108000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:7341 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...d '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpp_er6idl/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:7341', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-14T11:26:34.896837Z :WARNING: No initial sequence number for ProducerId 6eac58f0-1f9811f1-b90b9076-ebe5ab7b PartitionId 39\nE   2026-03-14T11:26:34.912591Z :WARNING: No initial sequence number for ProducerId 6f9bae5a-1f9811f1-abbfbfb4-47a7715d PartitionId 12\nE   2026-03-14T11:26:34.919950Z :WARNING: No initial sequence number for ProducerId 6fb3708a-1f9811f1-a4316323-6ddb71c7 PartitionId 13\nE   2026-03-14T11:26:34.931611Z :WARNING: No initial sequence number for ProducerId 6f9e98c2-1f9811f1-ae5ca8ee-9b6f27bd PartitionId 10\nE   2026-03-14T11:26:34.984745Z :WARNING: No initial sequence number for ProducerId 6eb53556-1f9811f1-aeba39e2-2da4274e PartitionId 44\nE   2026-03-14T11:26:34.984832Z :WARNING: No initial sequence number for ProducerId 6ea3c32a-1f9811f1-b2ea5329-e7628985 PartitionId 47\nE   2026-03-14T11:26:35.000658Z :WARNING: No initial sequence number for ProducerId 6ea0b482-1f9811f1-b2f73b8f-e3efab85 PartitionId 44\nE   2026-03-14T11:26:35.001201Z :WARNING: No initial sequence number for ProducerId 6fa07606-1f9811f1-84b651ba-e9ca664f PartitionId 13","1773485812000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:64647 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpb0necfi9/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:64647', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773485425000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13849 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...cerId 5125ea46-1f9111f1-a77ac580-19e5a4f0 PartitionId 1\nE   2026-03-14T10:35:22.366903Z :WARNING: No initial sequence number for ProducerId 515f73ce-1f9111f1-862f35dd-79d94958 PartitionId 14\nE   2026-03-14T10:35:22.385461Z :WARNING: No initial sequence number for ProducerId 5145eff8-1f9111f1-8769a896-874a9810 PartitionId 94\nE   2026-03-14T10:35:22.389885Z :WARNING: No initial sequence number for ProducerId 51429d58-1f9111f1-a6f08a36-e174bab6 PartitionId 97\nE   2026-03-14T10:35:22.421380Z :WARNING: No initial sequence number for ProducerId 516d0cbe-1f9111f1-a57eb403-2fcc7599 PartitionId 25\nE   2026-03-14T10:35:22.429694Z :WARNING: No initial sequence number for ProducerId 516e95fc-1f9111f1-9dbcd313-afbe8e7b PartitionId 19\nE   2026-03-14T10:35:22.437490Z :WARNING: No initial sequence number for ProducerId 51660a2c-1f9111f1-93c32c88-9beb7d09 PartitionId 18\nE   2026-03-14T10:35:22.454506Z :WARNING: No initial sequence number for ProducerId 524e5d36-1f9111f1-aabc5293-ebfabfa1 PartitionId 26\nE   2026-03-14T10:35:22.457487Z :WARNING: No initial sequence number for ProducerId 5174675c-1f9111f1-a8af06ab-eb4a613b PartitionId 25\nE   2026-03-14T10:35:22.478268Z :WARNING: No initial sequence number for ProducerId 519145ca-1f9111f1-87ea699d-850de722 PartitionId 21\nE   2026-03-14T10:35:22.478411Z :WARNING: No initial sequence number for ProducerId 517b4536-1f9111f1-a2ec2961-dd2a07a1 PartitionId 25\nE   2026-03-14T10:35:22.491151Z :WARNING: No initial sequence number for ProducerId 516e8e18-1f9111f1-966a5403-6d232bc5 PartitionId 20","1773485307000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:25601 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpffx4tzoy/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:25601', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773485102000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:63106 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpvrczi0py/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:63106', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773481064000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:19429 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...    '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpl2phn652/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:19429', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p10_pq10']' returned non-zero exit status 1.\nE   2026-03-14T09:16:35.425003Z :WARNING: No initial sequence number for ProducerId 532f72ea-1f8611f1-b83fe930-cb240fbe PartitionId 10\nE   No authentication methods were found. Going without authentication\nE   2026-03-14T09:16:35.472995Z :WARNING: No initial sequence number for ProducerId 532f272c-1f8611f1-8a29ea4d-8710d26e PartitionId 16\nE   2026-03-14T09:16:35.474589Z :WARNING: No initial sequence number for ProducerId 532db7c0-1f8611f1-83fb8d25-fdb05038 PartitionId 10\nE   2026-03-14T09:16:35.515459Z :WARNING: No initial sequence number for ProducerId 53290a68-1f8611f1-8fd8c44c-99580233 PartitionId 12","1773464419000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:7993 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ot', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpptwx4c_m/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:7993', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773464285000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:28966 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpdasuibwz/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:28966', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773463643000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:20025 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp0evzaiuo/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:20025', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773451079000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:8301 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ot', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp0bquz_kg/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:8301', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773447423000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:22513 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpty041eti/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:22513', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773428652000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:10243 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmplgyx8btw/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:10243', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773427126000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23489 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp5dom6kpl/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:23489', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773423092000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:15057 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp49bsff2u/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:15057', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773422300000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13647 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...cerId fb5c0ea2-1efe11f1-b5e03601-6ff1131f PartitionId 25\nE   2026-03-13T17:07:51.741424Z :WARNING: No initial sequence number for ProducerId fb553f78-1efe11f1-87fc9758-102e34b PartitionId 25\nE   2026-03-13T17:07:51.755581Z :WARNING: No initial sequence number for ProducerId fb35af14-1efe11f1-a43672fd-a12bf50c PartitionId 25\nE   2026-03-13T17:07:51.756903Z :WARNING: No initial sequence number for ProducerId fb9e03c0-1efe11f1-b13cb41f-5911fa82 PartitionId 25\nE   2026-03-13T17:07:51.758052Z :WARNING: No initial sequence number for ProducerId fb50a4f4-1efe11f1-847a3cc5-c5569969 PartitionId 25\nE   2026-03-13T17:07:51.758279Z :WARNING: No initial sequence number for ProducerId fb450086-1efe11f1-9b7c348b-a9ac2651 PartitionId 25\nE   2026-03-13T17:07:51.764668Z :WARNING: No initial sequence number for ProducerId fb5b8220-1efe11f1-a50b5923-ede20203 PartitionId 25\nE   2026-03-13T17:07:51.795239Z :WARNING: No initial sequence number for ProducerId fadcf7fc-1efe11f1-aae5cc82-39795a03 PartitionId 24\nE   2026-03-13T17:07:51.797057Z :WARNING: No initial sequence number for ProducerId fb504d06-1efe11f1-be33d024-790430dd PartitionId 25\nE   2026-03-13T17:07:51.797126Z :WARNING: No initial sequence number for ProducerId fb581a0e-1efe11f1-860b75a5-45d4987a PartitionId 25\nE   2026-03-13T17:07:51.811355Z :WARNING: No initial sequence number for ProducerId fb5ceffc-1efe11f1-bea3505d-7d2557af PartitionId 25\nE   2026-03-13T17:07:51.811650Z :WARNING: No initial sequence number for ProducerId fb781d0e-1efe11f1-a118c8d9-7f12aeb3 PartitionId 25","1773421818000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23854 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ber for ProducerId d18bd5e-1efd11f1-9982dcd9-51be8bc0 PartitionId 70\nE   2026-03-13T16:54:05.421395Z :WARNING: No initial sequence number for ProducerId d1a2f54-1efd11f1-b38246ea-a5b0707f PartitionId 66\nE   2026-03-13T16:54:05.421423Z :WARNING: No initial sequence number for ProducerId d17e442-1efd11f1-b18fb22c-61d8f57b PartitionId 76\nE   2026-03-13T16:54:05.421451Z :WARNING: No initial sequence number for ProducerId d3d263a-1efd11f1-adc6c99b-77739db4 PartitionId 78\nE   2026-03-13T16:54:05.432053Z :WARNING: No initial sequence number for ProducerId dada4b4-1efd11f1-b3bf282c-69810b37 PartitionId 11\nE   2026-03-13T16:54:05.432050Z :WARNING: No initial sequence number for ProducerId d7682c2-1efd11f1-921438d6-d364a41a PartitionId 6\nE   2026-03-13T16:54:05.433665Z :WARNING: No initial sequence number for ProducerId d15acb8-1efd11f1-ac066980-47e1d7e4 PartitionId 73\nE   2026-03-13T16:54:05.430729Z :WARNING: No initial sequence number for ProducerId d3b4e00-1efd11f1-a876ef8c-cf12941c PartitionId 76\nE   2026-03-13T16:54:05.431227Z :WARNING: No initial sequence number for ProducerId d184504-1efd11f1-a09f95a4-f189518 PartitionId 66\nE   2026-03-13T16:54:05.424513Z :WARNING: No initial sequence number for ProducerId cf8aece-1efd11f1-93d3729f-a98b033c PartitionId 80\nE   2026-03-13T16:54:05.433733Z :WARNING: No initial sequence number for ProducerId da279d6-1efd11f1-a93dece6-2b55244c PartitionId 12\nE   2026-03-13T16:54:05.436657Z :WARNING: No initial sequence number for ProducerId daf091c-1efd11f1-b61b3980-7b2ae5b1 PartitionId 10","1773421534000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:5562 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...q10']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 71, in __two_tablets_distributed_transaction\nE       self.run_topic_write_with_tx(20, 5, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpuz6kfg4q/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:5562', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p5_pq10']' returned non-zero exit status 1.","1773419832000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:63674 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp82buzm5u/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:63674', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773419396000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:4704 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp0uzvgy1z/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:4704', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773419081000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13726 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp98bm7c33/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:13726', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773418484000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:21867 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...s/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp0en9lw2k/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:21867', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-13T16:06:27.376850Z :EMERG: No messages were written.","1773418466000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:29687 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmph4khhcby/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:29687', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773418209000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:30176 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp2b7hsl7_/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:30176', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773417840000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:30754 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpqt_1usi9/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:30754', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773417667000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:25098 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpveu588nk/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:25098', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773417296000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:12919 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmprg6_b_7s/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:12919', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773417175000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:4766 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...f.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp8pflkj5n/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:4766', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-13T15:45:22.684501Z :WARNING: No initial sequence number for ProducerId 733c10b8-1ef311f1-9aa021b8-1daa1508 PartitionId 66\nE   2026-03-13T15:45:22.743625Z :WARNING: No initial sequence number for ProducerId 7338cd9a-1ef311f1-bbcab23b-93d22bb PartitionId 64","1773417033000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:6232 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp7dnqwqcj/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:6232', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773416973000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:20042 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpyk8n99h6/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:20042', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773416913000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:3186 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...kload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpjf8d51u_/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:3186', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773416723000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:28920 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...roducerId c7827fa0-1ef211f1-9c678382-6fbc4e89 PartitionId 8\nE   2026-03-13T15:40:37.446317Z :WARNING: No initial sequence number for ProducerId c7759db2-1ef211f1-84826a49-a59c1b58 PartitionId 8\nE   2026-03-13T15:40:37.450143Z :WARNING: No initial sequence number for ProducerId c665aa84-1ef211f1-85578fac-b1a31d24 PartitionId 13\nE   2026-03-13T15:40:37.450142Z :WARNING: No initial sequence number for ProducerId c776a2f2-1ef211f1-89cee34c-3954e80d PartitionId 7\nE   2026-03-13T15:40:37.450372Z :WARNING: No initial sequence number for ProducerId c664fdf0-1ef211f1-8562b8e7-17f64572 PartitionId 13\nE   2026-03-13T15:40:37.450812Z :WARNING: No initial sequence number for ProducerId c60d014a-1ef211f1-b02472ce-758645fc PartitionId 86\nE   2026-03-13T15:40:37.451627Z :WARNING: No initial sequence number for ProducerId c5c486d6-1ef211f1-9cec5a1b-8747d9d1 PartitionId 84\nE   2026-03-13T15:40:37.454172Z :WARNING: No initial sequence number for ProducerId c662c4a4-1ef211f1-81704c31-7bc1455d PartitionId 15\nE   2026-03-13T15:40:37.455010Z :WARNING: No initial sequence number for ProducerId c615b74a-1ef211f1-a82f0251-d17c563a PartitionId 91\nE   2026-03-13T15:40:37.458677Z :WARNING: No initial sequence number for ProducerId c65fd0d2-1ef211f1-b844f513-f1f9ccea PartitionId 8\nE   2026-03-13T15:40:37.458926Z :WARNING: No initial sequence number for ProducerId c5b06e30-1ef211f1-a7d239e1-7fb6910 PartitionId 77\nE   2026-03-13T15:40:37.462635Z :WARNING: No initial sequence number for ProducerId c621eede-1ef211f1-80ebadc8-c723f943 PartitionId 91","1773416600000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:25031 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpet62_g15/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:25031', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773416491000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:65168 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpln2gbewr/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:65168', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T15:34:48.859709Z :WARNING: No initial sequence number for ProducerId fde16990-1ef111f1-b4afc442-d52c183d PartitionId 52\nE   2026-03-13T15:34:48.863179Z :WARNING: No initial sequence number for ProducerId fde81df8-1ef111f1-ab7ad38a-291003f6 PartitionId 57\nE   2026-03-13T15:34:48.863448Z :WARNING: No initial sequence number for ProducerId fe2408c2-1ef111f1-81781f74-63ca30cc PartitionId 58\nE   2026-03-13T15:34:48.863383Z :WARNING: No initial sequence number for ProducerId ff199f44-1ef111f1-883ea8f5-63149d30 PartitionId 17\nE   2026-03-13T15:34:48.871508Z :WARNING: No initial sequence number for ProducerId fdd7e6ea-1ef111f1-a51a8d4f-d1094ba9 PartitionId 54\nE   2026-03-13T15:34:48.871509Z :WARNING: No initial sequence number for ProducerId fdf11502-1ef111f1-8da67377-75ed2a66 PartitionId 53\nE   2026-03-13T15:34:48.871985Z :WARNING: No initial sequence number for ProducerId fe813a92-1ef111f1-9175deec-fbeef343 PartitionId 65\nE   2026-03-13T15:34:48.875597Z :WARNING: No initial sequence number for ProducerId fe087ea4-1ef111f1-843836fc-ffd49b23 PartitionId 61","1773414132000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:18043 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ducerId 48b76664-1eec11f1-b22cd053-1bd7c486 PartitionId 56\nE   2026-03-13T14:53:54.220708Z :WARNING: No initial sequence number for ProducerId 49580ace-1eec11f1-b0d31369-357c3d7f PartitionId 28\nE   2026-03-13T14:53:54.222902Z :WARNING: No initial sequence number for ProducerId 4958fe70-1eec11f1-afbee98c-319cb5e0 PartitionId 23\nE   2026-03-13T14:53:54.225934Z :WARNING: No initial sequence number for ProducerId 4965a8be-1eec11f1-be858ed1-d99c68b7 PartitionId 26\nE   2026-03-13T14:53:54.227669Z :WARNING: No initial sequence number for ProducerId 48daf0e8-1eec11f1-a0fb3b17-2bdf335e PartitionId 54\nE   2026-03-13T14:53:54.229085Z :WARNING: No initial sequence number for ProducerId 494c4e96-1eec11f1-b9943e38-b3851d7f PartitionId 23\nE   2026-03-13T14:53:54.229221Z :WARNING: No initial sequence number for ProducerId 481a61b6-1eec11f1-b54e41bc-ade882b6 PartitionId 0\nE   2026-03-13T14:53:54.229316Z :WARNING: No initial sequence number for ProducerId 482c9cf0-1eec11f1-b77a7637-b7dee5f7 PartitionId 2\nE   2026-03-13T14:53:54.229355Z :WARNING: No initial sequence number for ProducerId 48ccf1d2-1eec11f1-912ed2e3-1be0ebdc PartitionId 51\nE   2026-03-13T14:53:54.232054Z :WARNING: No initial sequence number for ProducerId 495a467c-1eec11f1-81273660-83935625 PartitionId 22\nE   2026-03-13T14:53:54.232144Z :WARNING: No initial sequence number for ProducerId 495ba792-1eec11f1-92214e66-c13e28b7 PartitionId 22\nE   2026-03-13T14:53:54.233985Z :WARNING: No initial sequence number for ProducerId 4829be54-1eec11f1-ac2d63f1-9f76b6fe PartitionId 2","1773413019000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:5325 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpffrzeoiu/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:5325', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773412901000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:6069 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...erId 7ab5d0d6-1ee911f1-b20cc1b2-71c300bb PartitionId 49\nE   2026-03-13T14:34:01.597757Z :WARNING: No initial sequence number for ProducerId 7b9b4224-1ee911f1-b1ba6e84-51aa9184 PartitionId 32\nE   2026-03-13T14:34:01.605683Z :WARNING: No initial sequence number for ProducerId 7b9aacba-1ee911f1-b2748bf7-914a9bd1 PartitionId 31\nE   2026-03-13T14:34:01.622837Z :WARNING: No initial sequence number for ProducerId 7b93c6f2-1ee911f1-9e4fad03-cb45d0ce PartitionId 31\nE   2026-03-13T14:34:01.633246Z :WARNING: No initial sequence number for ProducerId 7b7df46c-1ee911f1-af2fae1d-cbc7eb67 PartitionId 29\nE   2026-03-13T14:34:01.663085Z :WARNING: No initial sequence number for ProducerId 7ade25e0-1ee911f1-96409ae6-5dacd2c9 PartitionId 50\nE   2026-03-13T14:34:01.670324Z :WARNING: No initial sequence number for ProducerId 7baa8ab8-1ee911f1-8f2652dd-cd60e5d5 PartitionId 30\nE   2026-03-13T14:34:01.671759Z :WARNING: No initial sequence number for ProducerId 7b844d6c-1ee911f1-bc4804b8-67c9af5c PartitionId 28\nE   2026-03-13T14:34:01.690779Z :WARNING: No initial sequence number for ProducerId 7bc5841c-1ee911f1-887d223e-7f4a4e01 PartitionId 29\nE   2026-03-13T14:34:01.709641Z :WARNING: No initial sequence number for ProducerId 7b967fe6-1ee911f1-8def4041-ede367de PartitionId 30\nE   2026-03-13T14:34:01.721520Z :WARNING: No initial sequence number for ProducerId 7bcbdb3c-1ee911f1-acb73e7c-8d35d3cd PartitionId 32\nE   2026-03-13T14:34:01.729751Z :WARNING: No initial sequence number for ProducerId 7b939bbe-1ee911f1-a335c2b1-65becbf3 PartitionId 29","1773412877000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:20854 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpctz8ldh1/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:20854', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T14:30:07.959131Z :WARNING: No initial sequence number for ProducerId f05cd498-1ee811f1-b1c5cbdd-81861daf PartitionId 96\nE   2026-03-13T14:30:07.969996Z :WARNING: No initial sequence number for ProducerId f0743bf6-1ee811f1-a1b1594b-7f8e9893 PartitionId 90\nE   2026-03-13T14:30:07.985763Z :WARNING: No initial sequence number for ProducerId f063ae76-1ee811f1-a95fe8a5-fd691b1f PartitionId 96","1773412278000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:5975 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ion\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp3jd8d7er/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:5975', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T14:23:02.994802Z :WARNING: No initial sequence number for ProducerId f70b11d4-1ee711f1-b9a91fc3-29f8034c PartitionId 28\nE   2026-03-13T14:23:03.006799Z :WARNING: No initial sequence number for ProducerId f6c5cd18-1ee711f1-b67db13a-3bd5e065 PartitionId 28","1773409896000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:28924 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...n.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpwdcpzdww/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:28924', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T13:37:39.330994Z :WARNING: No initial sequence number for ProducerId 9e839c26-1ee111f1-9709f0aa-ed61a418 PartitionId 47","1773408162000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:25363 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpx6plryz1/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:25363', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773407266000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:8748 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...mon.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmplfmf6rs9/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:8748', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T13:02:00.870554Z :WARNING: No initial sequence number for ProducerId a21020ee-1edc11f1-84680a76-e79677d3 PartitionId 9","1773407147000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:24319 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...f.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmph3vbjw3f/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:24319', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-13T12:58:06.790895Z :WARNING: No initial sequence number for ProducerId ce9bdd6-1edc11f1-80b959be-85894a36 PartitionId 23\nE   2026-03-13T12:58:06.814457Z :WARNING: No initial sequence number for ProducerId ced49d8-1edc11f1-8a883413-d5247443 PartitionId 17","1773406369000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:64561 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... for ProducerId bcf95b26-1ed911f1-abed23c2-437c34d3 PartitionId 4\nE   2026-03-13T12:41:19.154715Z :WARNING: No initial sequence number for ProducerId bc7547d2-1ed911f1-96975042-a3d39a32 PartitionId 95\nE   2026-03-13T12:41:19.166177Z :WARNING: No initial sequence number for ProducerId bc8e978c-1ed911f1-97ab0149-5733469a PartitionId 98\nE   2026-03-13T12:41:19.166277Z :WARNING: No initial sequence number for ProducerId be0d96c6-1ed911f1-9f3cf7c0-fd189def PartitionId 4\nE   2026-03-13T12:41:19.176473Z :WARNING: No initial sequence number for ProducerId be0c4a78-1ed911f1-830452ef-29be0feb PartitionId 3\nE   2026-03-13T12:41:19.177356Z :WARNING: No initial sequence number for ProducerId be05d59e-1ed911f1-bf1f61df-63fecb6f PartitionId 2\nE   2026-03-13T12:41:19.179745Z :WARNING: No initial sequence number for ProducerId bca377ce-1ed911f1-b0200f9a-9f722d3b PartitionId 2\nE   2026-03-13T12:41:19.185217Z :WARNING: No initial sequence number for ProducerId be134cce-1ed911f1-b6eb4b2b-5d2aaa2 PartitionId 9\nE   2026-03-13T12:41:19.185355Z :WARNING: No initial sequence number for ProducerId be05cbd0-1ed911f1-af147eba-4deba637 PartitionId 2\nE   2026-03-13T12:41:19.185312Z :WARNING: No initial sequence number for ProducerId bd00c7f8-1ed911f1-95eb23d5-d5c1d3c5 PartitionId 4\nE   2026-03-13T12:41:19.185434Z :WARNING: No initial sequence number for ProducerId be0d6bc4-1ed911f1-8305d269-33cc904b PartitionId 5\nE   2026-03-13T12:41:19.186209Z :WARNING: No initial sequence number for ProducerId bd00b592-1ed911f1-90f5bb06-ff903f22 PartitionId 5","1773405522000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:9068 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...mon.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp2633enix/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:9068', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T12:29:21.761545Z :WARNING: No initial sequence number for ProducerId a349e0c-1ed811f1-91d5e581-51313327 PartitionId 29","1773404601000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:27430 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp6hb7zp1g/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:27430', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773402640000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:61894 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpwyhhp26s/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:61894', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773401100000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:24822 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t__.py\", line 68, in __one_tablet_but_a_distributed_transaction\nE       self.run_topic_write_with_tx(20, 10, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpe9bes247/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:24822', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p10_pq10']' returned non-zero exit status 1.\nE   2026-03-13T11:19:44.480351Z :WARNING: No initial sequence number for ProducerId 5a5b8dd2-1ece11f1-b40b3233-ad4a84c2 PartitionId 2\nE   2026-03-13T11:19:44.526136Z :EMERG: No messages were written.","1773398367000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23003 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmptacrcjb3/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:23003', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p10_pq10']' returned non-zero exit status 1.\nE   2026-03-13T10:27:18.918086Z :WARNING: No initial sequence number for ProducerId fe0800f8-1ec611f1-9df1b587-df4d2d25 PartitionId 19\nE   2026-03-13T10:27:18.927911Z :WARNING: No initial sequence number for ProducerId fe12984c-1ec611f1-95d58245-65b2f943 PartitionId 16\nE   2026-03-13T10:27:19.033120Z :WARNING: No initial sequence number for ProducerId fe32d21a-1ec611f1-a1ff5e7f-9962e5b5 PartitionId 15","1773397572000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:29319 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp9_zzg_sl/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:29319', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.","1773397083000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:20722 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...lf.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp6m7dsn2l/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:20722', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1.\nE   2026-03-13T10:03:54.428394Z :WARNING: No initial sequence number for ProducerId ba95682c-1ec311f1-b6d956c9-fdaf1947 PartitionId 71\nE   2026-03-13T10:03:54.428483Z :WARNING: No initial sequence number for ProducerId ba4b82b6-1ec311f1-965f48ac-292535d3 PartitionId 65","1773394916000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23151 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpw5lkiz1_/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:23151', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T09:33:49.048165Z :EMERG: No messages were written.","1773394033000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:24893 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpygb3k8ck/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:24893', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773392201000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:3198 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...rocessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp2tvvr1nt/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:3198', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T08:45:50.575729Z :WARNING: No initial sequence number for ProducerId d71f869a-1eb811f1-aa0f9270-9de8d830 PartitionId 20\nE   2026-03-13T08:45:50.651020Z :WARNING: No initial sequence number for ProducerId d730dce2-1eb811f1-9230afb6-d5268926 PartitionId 19\nE   2026-03-13T08:45:50.654166Z :WARNING: No initial sequence number for ProducerId d5fcdae2-1eb811f1-bb38994a-89f9d8ac PartitionId 72\nE   2026-03-13T08:45:50.680241Z :WARNING: No initial sequence number for ProducerId d60dfcf0-1eb811f1-b698d3d8-5b48150d PartitionId 85\nE   2026-03-13T08:45:50.681745Z :WARNING: No initial sequence number for ProducerId d733d7d0-1eb811f1-bab64817-11f53da0 PartitionId 21\nE   2026-03-13T08:45:50.707216Z :WARNING: No initial sequence number for ProducerId d708606e-1eb811f1-ab6a2bab-d3962776 PartitionId 21\nE   2026-03-13T08:45:50.707326Z :WARNING: No initial sequence number for ProducerId d5ef2fb4-1eb811f1-af76c0d3-db3297dd PartitionId 72","1773391749000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:23724 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ucerId 72731046-1eb711f1-bc8c5a7e-fb333cb1 PartitionId 30\nE   2026-03-13T08:35:56.786198Z :WARNING: No initial sequence number for ProducerId 727390ac-1eb711f1-bda58721-997a8b56 PartitionId 29\nE   2026-03-13T08:35:56.786235Z :WARNING: No initial sequence number for ProducerId 727bee14-1eb711f1-9d1aaad2-43b81c7c PartitionId 30\nE   2026-03-13T08:35:56.786267Z :WARNING: No initial sequence number for ProducerId 72716728-1eb711f1-87b22818-f3914b73 PartitionId 0\nE   2026-03-13T08:35:56.786315Z :WARNING: No initial sequence number for ProducerId 72683964-1eb711f1-80daaa02-ed42c8ae PartitionId 29\nE   2026-03-13T08:35:56.788952Z :WARNING: No initial sequence number for ProducerId 7273db34-1eb711f1-85b7a182-29c155ab PartitionId 30\nE   2026-03-13T08:35:56.814672Z :WARNING: No initial sequence number for ProducerId 72653b2e-1eb711f1-ba984b75-899d00a6 PartitionId 26\nE   2026-03-13T08:35:56.817289Z :WARNING: No initial sequence number for ProducerId 717b4136-1eb711f1-83cda03b-9123dbb PartitionId 48\nE   2026-03-13T08:35:56.823491Z :WARNING: No initial sequence number for ProducerId 7177e96e-1eb711f1-b7b25873-ff8ebe36 PartitionId 53\nE   2026-03-13T08:35:56.823543Z :WARNING: No initial sequence number for ProducerId 717d3c48-1eb711f1-bb89d78e-45181a7b PartitionId 50\nE   2026-03-13T08:35:56.898957Z :WARNING: No initial sequence number for ProducerId 7188c36a-1eb711f1-b3f5a6d3-ab1ec549 PartitionId 55\nE   2026-03-13T08:35:56.904183Z :WARNING: No initial sequence number for ProducerId 7191e86e-1eb711f1-af1a7ac5-edc64dcc PartitionId 48","1773391647000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:5623 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...f.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpng_ls940/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:5623', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.\nE   2026-03-13T08:39:33.879320Z :WARNING: No initial sequence number for ProducerId f0f36498-1eb711f1-9ac0def0-15199ee PartitionId 32\nE   2026-03-13T08:39:33.934258Z :WARNING: No initial sequence number for ProducerId f0f2f904-1eb711f1-a1760e0e-e78c80dc PartitionId 29","1773389368000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:30440 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpigqt05d6/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:30440', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773389248000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:17005 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpddmsgfgu/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:17005', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773389101000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:6937 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp_o3v9r87/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:6937', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773389088000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:27433 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...^^^^^^^^^^^^^\nE       ))\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 264, in run_topic_write_with_tx\nE       self._run_workload(\nE       ~~~~~~~~~~~~~~~~~~^\nE           topic_name,\nE           ^^^^^^^^^^^\nE       ...<6 lines>...\nE           with_config=True\nE           ^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 140, in _run_workload\nE       self.cmd_run_with_monitoring(self.get_command_prefix(subcmds=args))\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 153, in cmd_run_with_monitoring\nE       executor.run_with_monitoring(cmd)\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^\nE     File \"ydb/tests/stress/topic/workload/command_executor.py\", line 56, in run_with_monitoring\nE       raise subprocess.CalledProcessError(\nE       ...<3 lines>...\nE       )\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp5n_5qezp/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:27433', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--window=30', '--byte-rate', '100M', '-p', '33', '-c', '16', '--topic', 'workload_topic_pr33_p10_pq5', '-t', '16', '--use-tx', '--tx-commit-interval', '2000', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--configure-consumers', '500', '--describe-topic', '--describe-consumer', 'data_holder']' returned non-zero exit status 1.","1773388600000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:26968 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...    ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmplbcx00_o/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:26968', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T07:28:38.650477Z :WARNING: No initial sequence number for ProducerId 53ec794-1eae11f1-946d53c1-8b32d5df PartitionId 64\nE   2026-03-13T07:28:38.659111Z :WARNING: No initial sequence number for ProducerId 531f8ca-1eae11f1-98a08171-dbfcecf0 PartitionId 73\nE   2026-03-13T07:28:38.659996Z :WARNING: No initial sequence number for ProducerId 52d5086-1eae11f1-a01e3338-9fe2f7d8 PartitionId 72\nE   2026-03-13T07:28:38.662981Z :WARNING: No initial sequence number for ProducerId 532b8d2-1eae11f1-904907bd-af5e690e PartitionId 66\nE   2026-03-13T07:28:38.665260Z :WARNING: No initial sequence number for ProducerId 52f823e-1eae11f1-8813887a-fddc119d PartitionId 66","1773378027000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:9962 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...ot', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpskuzlz4i/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:9962', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","1773377371000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:13279 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...rgs(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp2lpghd_6/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:13279', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p5_pq10']' returned non-zero exit status 1.\nE   2026-03-13T04:39:25.645752Z :WARNING: No initial sequence number for ProducerId 6ad95618-1e9611f1-aec9c943-2bc0b688 PartitionId 14\nE   2026-03-13T04:39:25.781808Z :WARNING: No initial sequence number for ProducerId 6c43c0a6-1e9611f1-98921c79-b13ef1a8 PartitionId 26\nE   2026-03-13T04:39:25.802020Z :WARNING: No initial sequence number for ProducerId 6b576878-1e9611f1-a3beaa68-4372a86b PartitionId 68\nE   2026-03-13T04:39:25.803998Z :WARNING: No initial sequence number for ProducerId 6c675066-1e9611f1-80409056-70766cf PartitionId 26","1773376892000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:20272 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...on\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpq80e_bgk/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:20272', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   2026-03-13T04:31:40.859049Z :WARNING: No initial sequence number for ProducerId 5a5c9990-1e9511f1-90179d83-d5365da0 PartitionId 19\nE   2026-03-13T04:31:40.886399Z :WARNING: No initial sequence number for ProducerId 5a5afd92-1e9511f1-8fd26a69-67eb88bd PartitionId 21","1773374483000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:27044 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp0xqtlarn/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:27044', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773366191000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:7426 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...kload_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmp041aqy7x/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:7426', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773365943000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:64444 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...load_topic_pr20_p1_pq1']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 80, in __immediate_transaction\nE       self.run_topic_write_with_tx(20, 1, 1, \"10M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpg_pznajp/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:64444', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '10M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p1_pq1']' returned non-zero exit status 1.","1773364460000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:14473 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ... status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 74, in __a_wide_transaction_with_multiple_partitions_in_one_tablet\nE       self.run_topic_write_with_tx(20, 100, 10, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpemfr7vm5/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:14473', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq10']' returned non-zero exit status 1.","1773361816000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:32560 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   .../Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpyz0vkkuv/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:32560', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 1.","1773361143000000":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:3149 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...n-zero exit status 1.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 77, in __wide_transaction_one_tablet_contains_one_partition\nE       self.run_topic_write_with_tx(20, 100, 1, \"100M\")\nE       ~~~~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 121, in run_topic_write_with_tx\nE       self.cmd_run([\nE       ~~~~~~~~~~~~^^\nE           *self._get_cli_common_args(),\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<6 lines>...\nE           '--topic', topic_name\nE           ^^^^^^^^^^^^^^^^^^^^^\nE       ])\nE       ^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpu2tlc668/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:3149', '--database=/Root', 'workload', 'topic', 'run', 'write', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-t', '20', '--max-memory-usage-per-producer=2M', '--topic', 'workload_topic_pr20_p100_pq1']' returned non-zero exit status 1."},"ydb/core/statistics/service/ut/unittest.[21/60] chunk":{"1773735610000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.05s","1773729526000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.87s","1773722559000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.33s","1773720290000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.43s","1773719674000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.22s","1773717082000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.20s","1773708708000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 599.48s","1773702149000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.29s","1773702129000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.07s","1773698398000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.31s","1773694511000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.44s","1773692039000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.58s","1773690853000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.98s","1773690441000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.78s","1773690098000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.30s","1773689559000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.98s","1773689025000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.46s","1773688476000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.66s","1773687448000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.33s","1773687172000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 599.55s","1773687167000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.58s","1773687158000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.36s","1773687146000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.43s","1773685622000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.25s","1773685521000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 605.38s","1773683098000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.92s","1773682225000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.13s","1773682040000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.99s","1773681818000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.94s","1773680250000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.96s","1773679933000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.36s","1773679449000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.94s","1773679120000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 600.82s","1773679007000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 605.24s","1773678560000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.11s","1773678224000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 599.92s","1773677952000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.80s","1773677718000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.29s","1773677510000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.96s","1773677230000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.85s","1773677073000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.84s","1773676288000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.08s","1773675427000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.36s","1773675081000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.31s","1773674966000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.36s","1773674896000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 609.80s","1773674709000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 605.07s","1773674432000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.49s","1773672978000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.86s","1773672747000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.74s","1773671187000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.92s","1773668553000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.42s","1773668400000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.79s","1773667338000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.86s","1773667038000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.85s","1773666791000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.76s","1773665446000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.86s","1773663833000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.58s","1773661068000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.42s","1773660879000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.41s","1773659987000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.89s","1773659879000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.02s","1773655409000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.49s","1773653768000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.38s","1773653670000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.99s","1773651160000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 605.26s","1773644268000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.21s","1773643731000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.86s","1773642105000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.14s","1773636520000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.38s","1773634122000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.31s","1773627098000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.37s","1773626829000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.33s","1773622583000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.15s","1773617923000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.26s","1773615765000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.88s","1773615308000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.85s","1773608284000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.71s","1773606561000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.32s","1773592886000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 605.37s","1773572070000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.84s","1773570977000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 605.37s","1773562367000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.40s","1773562210000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.40s","1773549981000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.94s","1773547892000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.40s","1773526956000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 600.06s","1773519197000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.70s","1773517997000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.50s","1773510184000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.38s","1773508856000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 605.83s","1773495254000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.20s","1773495175000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.51s","1773494895000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.33s","1773489674000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.90s","1773488352000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.88s","1773488108000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.83s","1773485812000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.41s","1773485425000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.74s","1773485307000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.35s","1773485102000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 599.19s","1773481064000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 608.83s","1773463643000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.30s","1773428652000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.55s","1773427126000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.19s","1773423092000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 608.83s","1773422300000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.76s","1773421818000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.60s","1773421534000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.41s","1773419832000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.95s","1773419396000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.45s","1773419081000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.88s","1773418484000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.13s","1773418466000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.00s","1773418209000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.92s","1773417840000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.54s","1773417667000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.43s","1773417296000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.87s","1773417033000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.33s","1773416973000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.16s","1773416913000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.35s","1773416723000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.80s","1773416600000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.51s","1773416491000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.88s","1773414132000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.78s","1773413019000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.15s","1773412901000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.77s","1773412877000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.82s","1773412278000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.77s","1773409896000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.74s","1773408162000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.11s","1773407266000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.82s","1773407147000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.60s","1773406369000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 605.30s","1773405522000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.99s","1773405071000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.83s","1773404601000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.67s","1773402640000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.87s","1773401100000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.04s","1773398367000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.47s","1773397572000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.66s","1773397083000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 611.78s","1773394916000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.71s","1773394033000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 608.85s","1773392201000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 605.12s","1773391749000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.30s","1773391647000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 614.05s","1773389491000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.28s","1773389368000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.51s","1773389248000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.38s","1773389101000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.55s","1773388600000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 609.88s","1773377371000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.37s","1773376892000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 602.97s","1773374483000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 609.88s","1773366191000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.61s","1773365943000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 600.26s","1773364460000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 603.35s","1773361816000000":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 604.08s"},"ydb/core/tx/sequenceproxy/ut/unittest.sole chunk":{"1773689025000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773677952000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)"},"ydb/core/graph/shard/ut/unittest.sole chunk":{"1773702129000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773690853000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773678560000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773677952000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773642105000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773606561000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773416973000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)"},"ydb/core/tx/sequenceshard/ut/unittest.sole chunk":{"1773678541000000":"4 tests: 3 - NOT_LAUNCHED, 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\nSequenceShardTests::Basics (timeout) duration: 35.41s\n3 tests were not launched inside chunk."},"ydb/core/tx/schemeshard/ut_truncate_table_simple/unittest.sole chunk":{"1773719674000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773708708000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773692039000000":"8 tests: 7 - GOOD, 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\nTruncateTable::TruncateTableWithSecondaryIndex (good) duration: 16.70s\nTruncateTable::TruncateTableWithUniqueIndex (timeout) duration: 8.73s\nTruncateTable::TruncateTableWithIndexAndCdcStream (good) duration: 8.38s\nTruncateTable::TruncateTableSequentialOperations (good) duration: 7.03s\nTruncateTable::TruncateNonExistentTable (good) duration: 7.00s\nTruncateTable::TruncateTableWithConcurrentTruncate (good) duration: 3.35s\nTruncateTable::TruncateTableWithCdcStream (good) duration: 1.96s\nTruncateTable::TruncateTableWithConcurrentDrop (good) duration: 1.55s","1773679120000000":"8 tests: 6 - GOOD, 1 - NOT_LAUNCHED, 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\nTruncateTable::TruncateTableWithSecondaryIndex (timeout) duration: 10.85s\nTruncateTable::TruncateTableWithConcurrentDrop (good) duration: 10.32s\nTruncateTable::TruncateTableWithCdcStream (good) duration: 10.27s\nTruncateTable::TruncateNonExistentTable (good) duration: 7.34s\nTruncateTable::TruncateTableSequentialOperations (good) duration: 5.10s\nTruncateTable::TruncateTableWithIndexAndCdcStream (good) duration: 4.60s\nTruncateTable::TruncateTableWithConcurrentTruncate (good) duration: 1.33s\nTruncateTable::TruncateTableWithUniqueIndex test was not launched inside chunk.","1773678560000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773678541000000":"8 tests: 7 - NOT_LAUNCHED, 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\nTruncateTable::TruncateTableWithConcurrentDrop (timeout) duration: 41.76s\n7 tests were not launched inside chunk.","1773671187000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773667038000000":"8 tests: 6 - GOOD, 1 - NOT_LAUNCHED, 1 - TIMEOUT\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\nTruncateTable::TruncateTableWithSecondaryIndex (timeout) duration: 24.74s\nTruncateTable::TruncateTableWithIndexAndCdcStream (good) duration: 13.17s\nTruncateTable::TruncateTableSequentialOperations (good) duration: 9.74s\nTruncateTable::TruncateTableWithConcurrentDrop (good) duration: 9.60s\nTruncateTable::TruncateNonExistentTable (good) duration: 9.44s\nTruncateTable::TruncateTableWithConcurrentTruncate (good) duration: 9.06s\nTruncateTable::TruncateTableWithCdcStream (good) duration: 2.92s\nTruncateTable::TruncateTableWithUniqueIndex test was not launched inside chunk.","1773666791000000":"8 tests: 7 - GOOD, 1 - TIMEOUT\n\nChunk exceeded 60s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\nTruncateTable::TruncateTableWithUniqueIndex (timeout) duration: 34.44s\nTruncateTable::TruncateTableWithConcurrentDrop (good) duration: 12.95s\nTruncateTable::TruncateTableWithSecondaryIndex (good) duration: 12.25s\nTruncateTable::TruncateTableWithIndexAndCdcStream (good) duration: 11.05s\nTruncateTable::TruncateTableWithConcurrentTruncate (good) duration: 6.51s\nTruncateTable::TruncateTableWithCdcStream (good) duration: 2.51s\nTruncateTable::TruncateNonExistentTable (good) duration: 1.94s\nTruncateTable::TruncateTableSequentialOperations (good) duration: 1.53s","1773659879000000":"8 tests: 5 - GOOD, 2 - NOT_LAUNCHED, 1 - TIMEOUT\n\nChunk exceeded 60s timeout, failed to shutdown gracefully in 30s and was terminated using SIGQUIT signal\nList of the tests involved in the launch:\nTruncateTable::TruncateTableWithIndexAndCdcStream (timeout) duration: 45.74s\nTruncateTable::TruncateTableWithConcurrentTruncate (good) duration: 12.98s\nTruncateTable::TruncateTableSequentialOperations (good) duration: 9.26s\nTruncateTable::TruncateTableWithConcurrentDrop (good) duration: 8.60s\nTruncateTable::TruncateTableWithCdcStream (good) duration: 6.82s\nTruncateTable::TruncateNonExistentTable (good) duration: 1.90s\n2 tests were not launched inside chunk.","1773642105000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773606561000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773485425000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773418209000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773417296000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773412901000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)"},"ydb/core/ymq/actor/cloud_events/cloud_events_ut/unittest.sole chunk":{"1773717082000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773708708000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773702129000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773690853000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773688476000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773687448000000":"4 tests: 1 - GOOD, 3 - NOT_LAUNCHED\n\nTest was terminated by signal 3\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\nTCloudEventsProcessorTests::TestCreateCloudEventProcessor (good) duration: 35.17s\n3 tests were not launched inside chunk.\nKilled by timeout (60 s)","1773687167000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773681818000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773678541000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773677952000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773677230000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773675427000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773671187000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773667338000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773660879000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773485102000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773419396000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773418209000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773416973000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773416600000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","1773412901000000":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)"}},"test_descriptions":{"ydb/core/graph/shard/ut/unittest.sole chunk":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","ydb/core/http_proxy/ut/inside_ydb_ut/TestKinesisHttpProxy.TestListStreamConsumersWithToken":"Test crashed (return code: -6)\nSee logs for more info","ydb/core/kqp/ut/sysview/KqpSystemView.Sessions":"assertion failed at ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:208, virtual void NKikimr::NKqp::NTestSuiteKqpSystemView::TTestCaseSessions::Execute_(NUnitTest::TTestContext &): (result.IsSuccess()) <main>: Fatal: Execution, code: 1060\n    <main>: Fatal: util/generic/maybe.cpp:5: TMaybe is empty, value type: TBasicString<char, std::char_traits<char>>, code: 1\n\n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C355D7B\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C8519EB\n2. /tmp//-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:208: Execute_ @ 0x1BF0D74D\n3. /tmp//-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36: operator() @ 0x1BF76057\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36:1) &> @ 0x1BF76057\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36:1) &> @ 0x1BF76057\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36:1) &> @ 0x1BF76057\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BF76057\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BF76057\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C889599\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C889599\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C889599\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C858A87\n13. /tmp//-S/ydb/core/kqp/ut/sysview/kqp_sys_view_ut.cpp:36: Execute @ 0x1BF75385\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C85A23F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C8836AC\n16. ??:0: ?? @ 0x7F10E26C0D8F\n17. ??:0: ?? @ 0x7F10E26C0E3F\n18. ??:0: ?? @ 0x19442028\n","ydb/core/statistics/service/ut/ColumnStatistics.EqWidthHistogram":"assertion failed at ydb/core/statistics/service/ut/ut_column_statistics.cpp:206, virtual void NKikimr::NStat::NTestSuiteColumnStatistics::TTestCaseEqWidthHistogram::Execute_(NUnitTest::TTestContext &): (!responses.at(0).Success) \nTBackTrace::Capture() at /-S/util/system/backtrace.cpp:284:9\nGetCurrentTest at /-S/library/cpp/testing/unittest/registar.cpp:70:12\nUnRef at /-S/util/generic/ptr.h:637:13\noperator() at /-S/ydb/core/statistics/service/ut/ut_column_statistics.cpp:0:1\nTColoredProcessor::Run(std::__y1::function<void ()>, TBasicString<char, std::__y1::char_traits<char>> const&, char const*, bool) at /-S/library/cpp/testing/unittest/utmain.cpp:527:13\n~__value_func at /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:403:16\nUnRef at /-S/util/generic/ptr.h:637:13\nNUnitTest::TTestFactory::Execute() at /-S/library/cpp/testing/unittest/registar.cpp:0:19\nNUnitTest::RunMain(int, char**) at /-S/library/cpp/testing/unittest/utmain.cpp:0:44\n?? at ??:0:0\n?? at ??:0:0\n_start at ??:0:0\n","ydb/core/statistics/service/ut/ColumnStatistics.ManyColumns":"Killed by timeout (600 s)","ydb/core/statistics/service/ut/unittest.[21/60] chunk":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\nColumnStatistics::ManyColumns (timeout) duration: 601.39s","ydb/core/tx/schemeshard/ut_truncate_table_simple/unittest.sole chunk":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","ydb/core/tx/sequenceproxy/ut/unittest.sole chunk":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","ydb/core/tx/sequenceshard/ut/unittest.sole chunk":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","ydb/core/ymq/actor/cloud_events/cloud_events_ut/unittest.sole chunk":"0 tests\n\nCannot obtain list of ut tests in the allotted time ('ut --list-verbose' worked longer than 30s)","ydb/public/sdk/cpp/src/client/topic/ut/with_direct_read_ut/Describe.DescribePartitionPermissions":"equal assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202, virtual void NYdb::Dev::NTopic::NTests::NTestSuiteDescribe::TTestCaseDescribePartitionPermissions::Execute_(NUnitTest::TTestContext &): resultStatus == status \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C2008CB\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F0BCB\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:202: Execute_ @ 0x1BDC5C20\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: operator() @ 0x1BDCD8C7\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDCD8C7\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDCD8C7\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12:5) &> @ 0x1BDCD8C7\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BDCD8C7\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BDCD8C7\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C71F189\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C71F189\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C71F189\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6F7C67\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/describe_topic_ut.cpp:12: Execute @ 0x1BDCCBA7\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6F941F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C71929C\n16. ??:0: ?? @ 0x7F865EDE1D8F\n17. ??:0: ?? @ 0x7F865EDE1E3F\n18. ??:0: ?? @ 0x1960B028\n","ydb/public/sdk/cpp/src/client/topic/ut/with_direct_read_ut/TxUsage.Sinks_Oltp_WriteToTopic_3_Table":"assertion failed at ydb/public/sdk/cpp/src/client/topic/ut/ut_utils/txusage_fixture.cpp:1022, void NYdb::Dev::NTopic::NTests::NTxUsage::TFixture::TestWriteToTopic26(): (messages.size() == 3) failed: (0 != 3) \n0. /-S/util/system/backtrace.cpp:284: ?? @ 0x1C2008CB\n1. /tmp//-S/library/cpp/testing/unittest/registar.cpp:46: RaiseError @ 0x1C6F0BCB\n2. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/ut_utils/txusage_fixture.cpp:1022: TestWriteToTopic26 @ 0x502FBCE1\n3. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6: operator() @ 0x1BE3F467\n4. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:179: __invoke<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE3F467\n5. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:251: __call<(lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE3F467\n6. /-S/contrib/libs/cxxsupp/libcxx/include/__type_traits/invoke.h:273: __invoke_r<void, (lambda at /-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6:1) &> @ 0x1BE3F467\n7. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:167: operator() @ 0x1BE3F467\n8. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:319: operator() @ 0x1BE3F467\n9. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:436: operator() @ 0x1C71F189\n10. /-S/contrib/libs/cxxsupp/libcxx/include/__functional/function.h:995: operator() @ 0x1C71F189\n11. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:527: Run @ 0x1C71F189\n12. /tmp//-S/library/cpp/testing/unittest/registar.cpp:378: Run @ 0x1C6F7C67\n13. /tmp//-S/ydb/public/sdk/cpp/src/client/topic/ut/topic_to_table_ut.cpp:6: Execute @ 0x1BE3E795\n14. /tmp//-S/library/cpp/testing/unittest/registar.cpp:499: Execute @ 0x1C6F941F\n15. /tmp//-S/library/cpp/testing/unittest/utmain.cpp:899: RunMain @ 0x1C71929C\n16. ??:0: ?? @ 0x7FF883B3BD8F\n17. ??:0: ?? @ 0x7FF883B3BE3F\n18. ??:0: ?? @ 0x1960B028\n","ydb/tests/datashard/truncate/concurrency/py3test.[test_truncate_table_concurrency.py 0/10] chunk":"1 test: 1 - GOOD\n\nChunk exceeded 60s timeout and was killed\nList of the tests involved in the launch:\ntest_truncate_table_concurrency.py::TestTruncateTableConcurrency::test_truncate_with_concurrent_rw_operations (good) duration: 43.32s\nKilled by timeout (60 s)","ydb/tests/datashard/ttl/test_ttl.py.TestTTL.test_ttl[table_Date_1_UNIQUE_SYNC-pk_types35-all_types35-index35-Date-UNIQUE-SYNC]":"ydb/tests/datashard/ttl/test_ttl.py:355: in test_ttl\n    self.do_test_ttl(table_name, pk_types, all_types, index, ttl, unique, sync)\nydb/tests/datashard/ttl/test_ttl.py:64: in do_test_ttl\n    self.select(table_name, pk_types, all_types, index, dml)\nydb/tests/datashard/ttl/test_ttl.py:127: in select\n    self.create_select(table_name, pk_types, all_types, index, i, 0, dml)\nydb/tests/datashard/ttl/test_ttl.py:147: in create_select\n    assert (\nE   AssertionError: Expected 0 rows, error when deleting 6 lines, table table_Date_1_UNIQUE_SYNC\nE   assert (1 == 1 and 1 == 0)\nE    +  where 1 = len([{'count': 1}])\nE    +  and   1 = {'count': 1}.count","ydb/tests/functional/audit/test_canonical_records.py.test_dstool_evict_vdisk_grpc":"ydb/tests/functional/audit/test_canonical_records.py:373: in test_dstool_evict_vdisk_grpc\n    execute_dstool_grpc(ydb_cluster, TOKEN, ['vdisk', 'evict', '--vdisk-ids', vdisk_id, '--ignore-degraded-group-check', '--ignore-failure-model-group-check'])\nydb/tests/functional/audit/helpers.py:60: in execute_dstool_grpc\n    assert False, f'Command\\n{full_cmd}\\n finished with exit code {proc_result.exit_code}, stderr:\\n\\n{proc_result.std_err.decode(\"utf-8\")}\\n\\nstdout:\\n{proc_result.std_out.decode(\"utf-8\")}'\nE   AssertionError: Command\nE   ['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/apps/dstool/ydb-dstool', '--endpoint', 'grpc://localhost:23211', 'vdisk', 'evict', '--vdisk-ids', '[82000000:1:0:0:0]', '--ignore-degraded-group-check', '--ignore-failure-model-group-check']\nE    finished with exit code 1, stderr:\nE   \nE   error, Request has failed: \nE   Command {\nE     ReassignGroupDisk {\nE       GroupId: 2181038080\nE       GroupGeneration: 1\nE     }\nE   }\nE   IgnoreGroupFailModelChecks: true\nE   IgnoreDegradedGroupsChecks: true\nE   \nE   Status {\nE     Success: true\nE     ReassignedItem {\nE       VDiskId {\nE         GroupID: 2181038080\nE         GroupGeneration: 1\nE         Ring: 0\nE         Domain: 0\nE         VDisk: 0\nE       }\nE       From {\nE         NodeId: 1\nE         PDiskId: 1000\nE         VSlotId: 1000\nE       }\nE       To {\nE         NodeId: 1\nE         PDiskId: 1\nE         VSlotId: 1000\nE       }\nE       FromFqdn: \"localhost\"\nE       FromPath: \"SectorMap:2:64\"\nE       ToFqdn: \"localhost\"\nE       ToPath: \"SectorMap:1:64\"\nE     }\nE   }\nE   ErrorDescription: \"DisintegratedByExpectedStatus GroupIds# [2181038080]\"\nE   ConfigTxSeqNo: 6\nE   GroupsGetDisintegratedByExpectedStatus: 2181038080\nE   \nE   \nE   \nE   \nE   stdout:","ydb/tests/functional/statistics/py3test.[test_analyze.py 0/10] chunk":"1 test: 1 - TIMEOUT\n\nChunk exceeded 600s timeout and was killed\nList of the tests involved in the launch:\ntest_analyze.py::test_basic (timeout) duration: 614.56s","ydb/tests/functional/statistics/test_analyze.py.test_basic":"Killed by timeout (600 s)","ydb/tests/stress/kv/tests/test_workload.py.TestYdbKvWorkload.test[column]":"ydb/tests/stress/kv/tests/test_workload.py:16: in test\n    yatest.common.execute([\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/kv/workload_kv --endpoint grpc://localhost:17688 --database /Root --duration 60 --store_type column --kv_prefix column' has failed with code 1.\nE   Errors:\nE   ...tabase=/Root', 'workload', 'kv', 'run', 'mixed', '--seconds', '60', '--threads', '10', '--cols', '5', '--len', '200', '--int-cols', '2', '--key-cols', '3', '--path', 'column/kv_column']' died with <Signals.SIGABRT: 6>.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/kv/workload/__init__.py\", line 69, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=[\"run\", \"mixed\",\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       ...<5 lines>...\nE                                            \"--key-cols\", \"3\"])\nE                                            ^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/kv/workload/__init__.py\", line 53, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/kv/tests/test-results/py3test/tmp56qn6u15/kv_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:17688', '--database=/Root', 'workload', 'kv', 'run', 'mixed', '--seconds', '60', '--threads', '10', '--cols', '5', '--len', '200', '--int-cols', '2', '--key-cols', '3', '--path', 'column/kv_column']' died with <Signals.SIGABRT: 6>.","ydb/tests/stress/topic/tests/test_workload_topic.py.TestYdbTopicWorkload.test":"ydb/tests/stress/topic/tests/test_workload_topic.py:31: in test\n    yatest.common.execute(cmd_args)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:410: in wait\n    self._finalise(check_exit_code)\nlibrary/python/testing/yatest_common/yatest/common/process.py:421: in _finalise\n    raise ExecutionError(self)\nE   yatest.common.process.ExecutionError: Command '/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/workload_topic --endpoint grpc://localhost:15614 --database /Root --duration 60 --consumers 16 --producers 33 --limit-memory-usage' has failed with code 1.\nE   Errors:\nE   ...t', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.\nE   Traceback (most recent call last):\nE     File \"ydb/tests/stress/common/common.py\", line 114, in wrapper\nE       f()\nE       ~^^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 108, in __loop\nE       self.cmd_run(\nE       ~~~~~~~~~~~~^\nE           self.get_command_prefix(subcmds=run_cmd_args)\nE           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE       )\nE       ^\nE     File \"ydb/tests/stress/topic/workload/__init__.py\", line 64, in cmd_run\nE       subprocess.run(cmd, check=True, text=True)\nE       ~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nE     File \"contrib/tools/python3/Lib/subprocess.py\", line 577, in run\nE       raise CalledProcessError(retcode, process.args,\nE                                output=stdout, stderr=stderr)\nE   subprocess.CalledProcessError: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/stress/topic/tests/test-results/py3test/tmpj3un3hnj/topic_ydb_cli/ydb_cli', '--verbose', '--endpoint', 'grpc://localhost:15614', '--database=/Root', 'workload', 'topic', 'run', 'full', '-s', '60', '--byte-rate', '100M', '--use-tx', '--tx-commit-interval', '2000', '-p', '33', '-c', '16', '--max-memory-usage-per-consumer=2M', '--max-memory-usage-per-producer=2M', '--topic', 'topic/topic']' returned non-zero exit status 100.","ydb/tests/tools/pq_read/test/test_timeout.py.TestTimeout.test_timeout":"library/python/testing/yatest_common/yatest/common/process.py:383: in wait\n    wait_for(\nlibrary/python/testing/yatest_common/yatest/common/process.py:764: in wait_for\n    raise TimeoutError(truncate(message, MAX_MESSAGE_LEN))\nE   yatest.common.process.TimeoutError: 6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:25331', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\n\nDuring handling of the above exception, another exception occurred:\nydb/tests/tools/pq_read/test/test_timeout.py:17: in test_timeout\n    assert read_stream(topic, len(data) + 42, commit_after_processing=True, consumer_name=consumer_name, timeout=3) == data\nydb/tests/tools/datastreams_helpers/data_plane.py:73: in read_stream\n    yatest.common.execute(cmd, timeout=timeout * 2, stdout=outfile)\nlibrary/python/testing/yatest_common/yatest/common/process.py:655: in execute\n    res.wait(check_exit_code, timeout, on_timeout)\nlibrary/python/testing/yatest_common/yatest/common/process.py:397: in wait\n    raise ExecutionTimeoutError(self, str(e))\nE   yatest.common.process.ExecutionTimeoutError: ((\"6 second(s) wait timeout has expired: Command '['/home/runner/actions_runner/_work/ydb/ydb/tmp/out/ydb/tests/tools/pq_read/pq_read', '--endpoint', 'localhost:25331', '--database', 'local', '--topic-path', 'timeout', '--consumer-name', 'test_client', '--disable-cluster-discovery', '--messages-count', '44', '--timeout', '3000ms', '--commit-after-processing']' stopped by 6 seconds timeout\",), {})","yql/essentials/core/cbo/cbo_optimizer_new.cpp.clang_format":"--- L\n+++ R\n@@ -378,7 +378,7 @@\n \n     auto result = TOptimizerStatistics(outputType, newCard, newNCols, newByteSize, cost,\n                                        leftKeyColumns ? leftStats.KeyColumns : (rightKeyColumns ? rightStats.KeyColumns : TIntrusivePtr<TOptimizerStatistics::TKeyColumns>()));\n-    //result.Selectivity = std::pow(selectivity, 0.7);\n+    // result.Selectivity = std::pow(selectivity, 0.7);\n     result.Selectivity = selectivity;\n     return result;\n }","ydb/tests/example/test_example.py.TestExample.test_skipped_with_issue":"Skipped: #999999999","ydb/tests/fq/solomon/test.py.test[solomon-Subquery-default.txt]":"Skipped: Using of system 'kikimr' is not allowed in SUBQUERY","ydb/tests/functional/api/test_discovery.py.TestDiscoveryFaultInjectionSlotStop.test_scenario":"ydb/tests/functional/api/test_discovery.py:198: in test_scenario\n    session.prepare(\ncontrib/python/ydb/py3/ydb/table.py:1815: in prepare\n    return self._driver(\ncontrib/python/ydb/py3/ydb/tracing.py:77: in wrapper\n    return f(self, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/pool.py:475: in __call__\n    res = connection(\ncontrib/python/ydb/py3/ydb/connection.py:542: in __call__\n    return response if wrap_result is None else wrap_result(rpc_state, response, *wrap_args)\ncontrib/python/ydb/py3/ydb/_session_impl.py:20: in decorator\n    return func(rpc_state, response_pb, session_state, *args, **kwargs)\ncontrib/python/ydb/py3/ydb/_session_impl.py:31: in wrap_prepare_query_response\n    issues._process_response(response_pb.operation)\ncontrib/python/ydb/py3/ydb/issues.py:266: in _process_response\n    raise exc_class(_format_response(response_proto), response_proto.issues)\nE   ydb.issues.Unavailable: message: \"Failed to create default pool in database /Root/database\" severity: 1 issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry shard unavailable error, status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } issues { message: \"Retry limit exceeded on status: StatusNotAvailable\" severity: 1 issues { message: \"Schemeshard not available\" severity: 1 } } ,message: \"Query failed during adding/waiting in workload pool \" severity: 1 (server_code: 400050)","ydb/tests/functional/nbs/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed":"ydb/tests/functional/nbs/test_nbs_load_actor.py:63: in test_nbs_load_actor_mixed\n    self.verify_load_test_results(results, expected_blocks_read=True, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:312: in verify_load_test_results\n    assert results['Result'] == 0, \"Load actor run finished with error\"\nE   AssertionError: Load actor run finished with error","ydb/tests/functional/nbs/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read":"ydb/tests/functional/nbs/test_nbs_load_actor.py:78: in test_nbs_load_actor_write_then_read\n    self.verify_load_test_results(write_results, expected_blocks_read=False, expected_blocks_written=True)\nydb/tests/functional/nbs/common.py:319: in verify_load_test_results\n    assert 'ThroughputMbs' in results, f\"Missing ThroughputMbs in results: {results}\"\nE   AssertionError: Missing ThroughputMbs in results: {'BlocksWritten': 3001, 'EndTime': 1773731020446554, 'Iops': 272, 'RequestsCompleted': 3001, 'StartTime': 1773731009445527, 'WriteLatency': {'Max': 477439, 'Mean': 116162.9937, 'Min': 12664, 'P50': 103359, 'P90': 192255, 'P95': 230655, 'P99': 307455, 'P999': 382463, 'StdDeviation': 58374.38589}}","ydb/tests/functional/statistics/test_restarts.py.test_basic":"ydb/tests/functional/statistics/test_restarts.py:95: in test_basic\n    assert_that(wait_for(get_base_stats_response, timeout_seconds=5),\nE   AssertionError: base stats available after restart","ydb/tests/olap/scenario/test_alter_tiering.py.TestAlterTiering.test[many_tables]":"ydb/tests/olap/scenario/conftest.py:126: in test\n    raise errors[0]\nydb/tests/olap/scenario/conftest.py:85: in worker\n    self._test_suffix(local_ctx, suffix, codes, idx)\nydb/tests/olap/scenario/conftest.py:136: in _test_suffix\n    ctx.executable(self, ctx)\nydb/tests/olap/scenario/test_alter_tiering.py:362: in scenario_many_tables\n    threads.start_and_wait_all()\nydb/tests/olap/common/thread_helper.py:49: in start_and_wait_all\n    self.join_all()\nydb/tests/olap/common/thread_helper.py:45: in join_all\n    thread.join(timeout=timeout)\nydb/tests/olap/common/thread_helper.py:18: in join\n    raise self.exc\nydb/tests/olap/common/thread_helper.py:11: in run\n    self.ret = self._target(*self._args, **self._kwargs)\nydb/tests/olap/scenario/test_alter_tiering.py:248: in _loop_scan\n    sth.execute_query(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:507: in execute_query\n    return self._run_with_expected_status(\nydb/tests/olap/scenario/helpers/scenario_tests_helper.py:366: in _run_with_expected_status\n    pytest.fail(f'Unexpected status: must be in {repr(expected_status)}, but get {repr(error or status)}')\nE   Failed: Unexpected status: must be in {<StatusCode.SUCCESS: 400000>}, but get GenericError('message: \"Scan failed at tablet 72075186224037941, reason: task_error:Error reading blob range for columns: { Blob: DS:4294967295:[72075186224037941:1:2:255:1:6240:0] Offset: 432 Size: 232 }, error: cannot get blob: curlCode: 28, Timeout was reached, detailed error: , curlCode: 28, Timeout was reached, status: ERROR\" issue_code: 2013 severity: 1 (server_code: 400080)')"},"test_success_rates":{"ydb/core/graph/shard/ut/unittest.sole chunk":{"pr_check":{"rate":96.6,"passed":112,"total":116},"other":{"rate":93.2,"passed":41,"total":44}},"ydb/core/http_proxy/ut/inside_ydb_ut/TestKinesisHttpProxy.TestListStreamConsumersWithToken":{"pr_check":{"rate":100.0,"passed":120,"total":120},"other":{"rate":100.0,"passed":44,"total":44}},"ydb/core/kqp/ut/sysview/KqpSystemView.Sessions":{"pr_check":{"rate":98.3,"passed":117,"total":119},"other":{"rate":100.0,"passed":45,"total":45}},"ydb/core/statistics/service/ut/ColumnStatistics.EqWidthHistogram":{"pr_check":{"rate":96.6,"passed":112,"total":116},"other":{"rate":100.0,"passed":45,"total":45}},"ydb/core/statistics/service/ut/ColumnStatistics.ManyColumns":{"pr_check":{"rate":0.0,"passed":0,"total":116},"other":{"rate":9.8,"passed":5,"total":51}},"ydb/core/statistics/service/ut/unittest.[21/60] chunk":{"pr_check":{"rate":1.7,"passed":2,"total":116},"other":{"rate":11.8,"passed":6,"total":51}},"ydb/core/tx/schemeshard/ut_truncate_table_simple/unittest.sole chunk":{"pr_check":{"rate":87.2,"passed":102,"total":117},"other":{"rate":97.7,"passed":43,"total":44}},"ydb/core/tx/sequenceproxy/ut/unittest.sole chunk":{"pr_check":{"rate":98.3,"passed":114,"total":116},"other":{"rate":100.0,"passed":44,"total":44}},"ydb/core/tx/sequenceshard/ut/unittest.sole chunk":{"pr_check":{"rate":99.1,"passed":115,"total":116},"other":{"rate":100.0,"passed":44,"total":44}},"ydb/core/ymq/actor/cloud_events/cloud_events_ut/unittest.sole chunk":{"pr_check":{"rate":86.2,"passed":100,"total":116},"other":{"rate":88.6,"passed":39,"total":44}},"ydb/public/sdk/cpp/src/client/topic/ut/with_direct_read_ut/Describe.DescribePartitionPermissions":{"pr_check":{"rate":90.5,"passed":105,"total":116},"other":{"rate":93.3,"passed":42,"total":45}},"ydb/public/sdk/cpp/src/client/topic/ut/with_direct_read_ut/TxUsage.Sinks_Oltp_WriteToTopic_3_Table":{"pr_check":{"rate":98.3,"passed":115,"total":117},"other":{"rate":97.8,"passed":44,"total":45}},"ydb/tests/datashard/truncate/concurrency/py3test.[test_truncate_table_concurrency.py 0/10] chunk":{"pr_check":{"rate":35.3,"passed":41,"total":116},"other":{"rate":41.9,"passed":18,"total":43}},"ydb/tests/datashard/ttl/test_ttl.py.TestTTL.test_ttl[table_Date_1_UNIQUE_SYNC-pk_types35-all_types35-index35-Date-UNIQUE-SYNC]":{"pr_check":{"rate":60.3,"passed":70,"total":116},"other":{"rate":75.6,"passed":34,"total":45}},"ydb/tests/functional/audit/test_canonical_records.py.test_dstool_evict_vdisk_grpc":{"pr_check":{"rate":0.0,"passed":0,"total":125},"other":{"rate":0.0,"passed":0,"total":53}},"ydb/tests/functional/statistics/py3test.[test_analyze.py 0/10] chunk":{"pr_check":{"rate":8.6,"passed":10,"total":116},"other":{"rate":17.0,"passed":8,"total":47}},"ydb/tests/functional/statistics/test_analyze.py.test_basic":{"pr_check":{"rate":14.7,"passed":17,"total":116},"other":{"rate":25.5,"passed":12,"total":47}},"ydb/tests/stress/kv/tests/test_workload.py.TestYdbKvWorkload.test[column]":{"pr_check":{"rate":98.3,"passed":117,"total":119},"other":{"rate":98.1,"passed":52,"total":53}},"ydb/tests/stress/topic/tests/test_workload_topic.py.TestYdbTopicWorkload.test":{"pr_check":{"rate":3.4,"passed":4,"total":118},"other":{"rate":17.4,"passed":12,"total":69}},"ydb/tests/tools/pq_read/test/test_timeout.py.TestTimeout.test_timeout":{"pr_check":{"rate":83.9,"passed":99,"total":118},"other":{"rate":86.0,"passed":37,"total":43}},"yql/essentials/core/cbo/cbo_optimizer_new.cpp.clang_format":{"pr_check":{"rate":71.4,"passed":5,"total":7},"other":{"rate":100.0,"passed":7,"total":7}},"ydb/tests/functional/api/test_discovery.py.TestDiscoveryFaultInjectionSlotStop.test_scenario":{"pr_check":{"rate":66.4,"passed":81,"total":122},"other":{"rate":56.2,"passed":27,"total":48}},"ydb/tests/functional/nbs/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_mixed":{"pr_check":{"rate":85.7,"passed":102,"total":119},"other":{"rate":77.8,"passed":35,"total":45}},"ydb/tests/functional/nbs/test_nbs_load_actor.py.TestNbsLoadActor.test_nbs_load_actor_write_then_read":{"pr_check":{"rate":59.8,"passed":70,"total":117},"other":{"rate":63.4,"passed":26,"total":41}},"ydb/tests/functional/statistics/test_restarts.py.test_basic":{"pr_check":{"rate":0.0,"passed":0,"total":116},"other":{"rate":9.8,"passed":5,"total":51}},"ydb/tests/olap/scenario/test_alter_tiering.py.TestAlterTiering.test[many_tables]":{"pr_check":{"rate":71.4,"passed":90,"total":126},"other":{"rate":78.3,"passed":36,"total":46}}}}