id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1276668472
i cant run bot Seguí todos los pasos indicados y me sale desconectado el bot Please explain more and I can assist if possible.
gharchive/issue
2022-06-20T09:54:26
2025-04-01T06:36:51.429924
{ "authors": [ "ConnorSwis", "elrubenhd" ], "repo": "ConnorSwis/casino-bot", "url": "https://github.com/ConnorSwis/casino-bot/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
736905227
Fix number of hashes I noticed that the hash bytes were really low compared to @lithp 's and my numbers. Tracked it down and seems they were being counted only for one of the contracts. After this the overhead for 32 byte chunks are lower than 1 byte chunks but I only measured for one block and arity of 16 Thank you, that was a bad mistake. I'll merge this and build a bit more on it. However, after a few tests, it looks like the overhead numbers change but their proportions stay the same: from 8% for 1-byte chunks to 110% to 32-byte chunks.
gharchive/pull-request
2020-11-05T12:48:57
2025-04-01T06:36:51.431508
{ "authors": [ "hmijail", "s1na" ], "repo": "ConsenSys/code_merklization_of_traces", "url": "https://github.com/ConsenSys/code_merklization_of_traces/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1754073119
Use Snapshot X quest This PR adds a guide for Linea members to use Snapshot X Closing this one due to conflicts, new PR here
gharchive/pull-request
2023-06-13T05:34:10
2025-04-01T06:36:51.433166
{ "authors": [ "zzuziak" ], "repo": "ConsenSys/doc.zk-evm", "url": "https://github.com/ConsenSys/doc.zk-evm/pull/117", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
293248098
removes salt from serialize This makes the serialized vault more secure, from my point of view it makes no sense to store the salt in the wallets instance. as anyone with accesses to this has accesses to salt. Additionally, if you provide salt manually this can open the door for more interesting scenarios for unlocking the vault, for example, 2fa or 2-way response. Is there a reason why salt is part of the instance from what I see it should be stored separately or at least given an option to store in a more secure sense. @moejoe90 This is not a good idea IMO, the reason for the salt is only to avoid rainbow attacks: namely that an attacker precomputes derived keys, then is able to grab lots of encrypted wallets from malware infested computers and uses the precomputed derived keys to quickly test a lot of passwords. It should always be possible to use only the password to decrypt a serialized keystore and that would not be possible with this suggested change. I can understand your choice against rainbow attack, that is one way of looking into this. I still think storing salt in the instance is not the best option. in order to open the possibility of some 2fa without resorting to splitting the password and hashing it against some other 2fa text.
gharchive/pull-request
2018-01-31T17:36:19
2025-04-01T06:36:51.435429
{ "authors": [ "Moejoe90", "christianlundkvist" ], "repo": "ConsenSys/eth-lightwallet", "url": "https://github.com/ConsenSys/eth-lightwallet/pull/184", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1470163778
Restate Integer Types as Aliases? Currently, the various integer types (e.g. u16) are declared using newtype which makes them fully defined types in their own right. However, they could be declared using type only which would make them subtypes. In essence, this would avoid all the annoying casting that we currently do! A key question, however, is whether or not these types would still compile down to native Java types (e.g. u16 -> short, etc). So, the issue with this is that Dafny does not compile subtypes down to native types. Rather, it compiles them to BigInteger. An analysis of our code base suggests this refactoring makes sense. Some observations: u8 This is widely used to represent a byte of data. Therefore, we could defined a type byte for this purpose. u16/u32/u64/u128. These are only used in memory.dfy, bytes.dfy and int.dfy. They are essentially only used to simplify converting a u256 to/from seq<byte>. u160/u256 Obviously these are used widely and end up with many casts. Therefore, we could replace u16/u32/u64/u128 with e.g. word16/word32/word64/word128 or perhaps w16/w32/w64/w128 which would remain as newtype. Then, only u160 and u256 remain and changed to being subtypes.
gharchive/issue
2022-11-30T20:11:02
2025-04-01T06:36:51.439843
{ "authors": [ "DavePearce" ], "repo": "ConsenSys/evm-dafny", "url": "https://github.com/ConsenSys/evm-dafny/issues/479", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1088665805
why is https://rimble.consensys.design down ? i cant access the website https://rimble.consensys.design/ where is a proper documentation? Also interested in this Would be interested in this as well ✌️ It seems this project has been abandoned as last commit was 2 years ago. You can run a local instance of the Storyblock server following these steps: Close the repository Install the dependencies at the root of the project (using yarn or npm). Go into the /storyblock folder and install the dependencies (using yarn or npm). Run storyblock package script inside the /storyblock folder. Go to http://localhost:6006/ (it will also open automatically in most OS).
gharchive/issue
2021-12-25T22:53:23
2025-04-01T06:36:51.443974
{ "authors": [ "JahnKhan", "blendor", "enzoferey", "warp-space" ], "repo": "ConsenSysMesh/rimble-ui", "url": "https://github.com/ConsenSysMesh/rimble-ui/issues/457", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1883986554
feat: As an Issuer, I want to set the attester address via my Portal What does this PR do? Stops using tx.origin for the attester field of an Attestation on creation Create an "AutoIssuerPortal" to show how to create a Portal where the end-user is the issuer Related ticket Fixes #154 Type of change [ ] Chore [ ] Bug fix [X] New feature [ ] Documentation update Check list [] Unit tests for any smart contract change [ ] Contracts and functions are documented Codecov Report Merging #159 (72a9263) into dev (afbbd2b) will increase coverage by 0.15%. The diff coverage is 100.00%. @@ Coverage Diff @@ ## dev #159 +/- ## ========================================== + Coverage 96.62% 96.77% +0.15% ========================================== Files 8 8 Lines 148 155 +7 Branches 37 37 ========================================== + Hits 143 150 +7 Misses 4 4 Partials 1 1 Files Changed Coverage Δ src/AttestationRegistry.sol 97.05% <100.00%> (-0.09%) :arrow_down: src/PortalRegistry.sol 96.00% <100.00%> (ø) src/example/EASPortal.sol 100.00% <100.00%> (ø) src/interface/AbstractPortal.sol 96.00% <100.00%> (+1.55%) :arrow_up: :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
gharchive/pull-request
2023-09-06T13:10:05
2025-04-01T06:36:51.453932
{ "authors": [ "alainncls", "codecov-commenter" ], "repo": "Consensys/linea-attestation-registry", "url": "https://github.com/Consensys/linea-attestation-registry/pull/159", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2705855919
chore(deps): Update dependencies What does this PR do? Updates compromised dependencies Type of change [X] Chore [ ] Bug fix [ ] New feature [ ] Documentation update Check list [ ] Unit tests for any smart contract change [ ] Contracts and functions are documented Codecov Report All modified and coverable lines are covered by tests :white_check_mark: Project coverage is 48.59%. Comparing base (ef284ae) to head (6677329). Additional details and impacted files @@ Coverage Diff @@ ## dev #848 +/- ## ======================================= Coverage 48.59% 48.59% ======================================= Files 18 18 Lines 391 391 Branches 63 63 ======================================= Hits 190 190 Misses 198 198 Partials 3 3 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
gharchive/pull-request
2024-11-29T18:26:39
2025-04-01T06:36:51.459803
{ "authors": [ "alainncls", "codecov-commenter" ], "repo": "Consensys/linea-attestation-registry", "url": "https://github.com/Consensys/linea-attestation-registry/pull/848", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1449377573
UI issue with distribution functions on contact details line width does not change legend is empty I've removed the legend option as this does not make sense if there is a single curve anyway.
gharchive/issue
2022-11-15T08:31:13
2025-04-01T06:36:51.482795
{ "authors": [ "mcrot", "pastewka" ], "repo": "ContactEngineering/topobank", "url": "https://github.com/ContactEngineering/topobank/issues/922", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2354705367
Add classical channel Add classical channel Added the image to public folder, because the channel image from Wrapcast for this image is inaccessible
gharchive/pull-request
2024-06-15T09:53:51
2025-04-01T06:36:51.510953
{ "authors": [ "TahirHameed74" ], "repo": "Coop-Records/sonata", "url": "https://github.com/Coop-Records/sonata/pull/224", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1261168949
⚠️ Samsung has degraded performance In 9d2834e, Samsung (https://samsung.com) experienced degraded performance: HTTP code: 200 Response time: 2058 ms Resolved: Samsung performance has improved in f6da008.
gharchive/issue
2022-06-05T21:51:34
2025-04-01T06:36:51.513448
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-electronicbrands", "url": "https://github.com/CoopPlayzz/Webstatus-electronicbrands/issues/1324", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1346173543
⚠️ Samsung has degraded performance In e463e91, Samsung (https://samsung.com) experienced degraded performance: HTTP code: 200 Response time: 1603 ms Resolved: Samsung performance has improved in 474bb5b.
gharchive/issue
2022-08-22T10:25:53
2025-04-01T06:36:51.515866
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-electronicbrands", "url": "https://github.com/CoopPlayzz/Webstatus-electronicbrands/issues/2660", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1400228415
⚠️ Fortnite has degraded performance In b28162e, Fortnite (https://fortnite.com) experienced degraded performance: HTTP code: 200 Response time: 1705 ms Resolved: Fortnite performance has improved in 1743222.
gharchive/issue
2022-10-06T19:59:32
2025-04-01T06:36:51.518477
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-games", "url": "https://github.com/CoopPlayzz/Webstatus-games/issues/1855", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1437295531
⚠️ Fortnite has degraded performance In 4bf9559, Fortnite (https://fortnite.com) experienced degraded performance: HTTP code: 200 Response time: 1529 ms Resolved: Fortnite performance has improved in 4eaaa92.
gharchive/issue
2022-11-06T05:48:08
2025-04-01T06:36:51.520776
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-games", "url": "https://github.com/CoopPlayzz/Webstatus-games/issues/1995", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1466919155
⚠️ Fortnite has degraded performance In 9d25ddc, Fortnite (https://fortnite.com) experienced degraded performance: HTTP code: 200 Response time: 1627 ms Resolved: Fortnite performance has improved in dc1a4e7.
gharchive/issue
2022-11-28T18:59:58
2025-04-01T06:36:51.523116
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-games", "url": "https://github.com/CoopPlayzz/Webstatus-games/issues/2362", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
640362862
Report input errors to contract in canonicalize_address/humanize_address Before https://github.com/CosmWasm/cosmwasm/pull/431 the errors erported back to the contract were very strange. On the one hand, too many errors around Region handling were reported. On the other hand, errors from the backend implementation (FfiError::Other) are not reported and contain insufficient information. ~The only thing that made some sense was rejecting non-UTF8 human addresses as invalid input. But even this indicates a problem in the standard library, not the contract.~ See also // TODO: would be nice if do_canonicalize_address could differentiate between different errors // from Api.canonical_address and return INVALID_INPUT for those cases as well. let result = do_canonicalize_address(api, ctx, source_ptr2, dest_ptr); match result.unwrap_err() { VmError::FfiErr { source: FfiError::Other { .. }, } => {} err => panic!("Incorrect error returned: {:?}", err), }; Since we are probably getting and error string with useful information via FfiError::Other soon, I guess the best is to return an error Region pointer. I'm wondering about pushing the remaining error handling changes (also https://github.com/CosmWasm/go-cosmwasm/issues/73 and https://github.com/CosmWasm/cosmwasm/issues/308) to 1.0. Since we will have one more release pre-0.39-cosmos-sdk, we don't need to throw all these in there. Unless we can get them done in the next few days, I would rather cut 0.9 without them. (I will most likely punt https://github.com/CosmWasm/go-cosmwasm/issues/73) This can easily be done this week. Since it is a mayor ABI breakage, I'd like to have it in all upcoming testnets. Will release an alpha today without this fix, since it does not affect contract developers. Okay, fair enough. I will prepare a 0.9-alpha for go-cosmwasm and wasmd today as well (to help test cosmjs)
gharchive/issue
2020-06-17T11:21:42
2025-04-01T06:36:51.577234
{ "authors": [ "ethanfrey", "webmaster128" ], "repo": "CosmWasm/cosmwasm", "url": "https://github.com/CosmWasm/cosmwasm/issues/433", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
208375418
Jitter at animation end I've just updated our code to utilise the new Motion framework instead of the old one built in to Material. Using the Motion.rotation(angle:) function on a FabButton (as primary button of a Material Menu) works fine in terms of animating the button. However, we're noticing a flash of the initial animation state at the end of the rotation (GIF recording) menu.views.first?.animate(animation: Motion.rotation(angle: -90)) Any ideas what might cause this? Can you try using motion for this? menu.views.first?.motion(.rotationAngle(-90)) Thanks! That fixed the issue. Awesome :)
gharchive/issue
2017-02-17T08:52:44
2025-04-01T06:36:51.579950
{ "authors": [ "danieldahan", "whitepixelstudios" ], "repo": "CosmicMind/Motion", "url": "https://github.com/CosmicMind/Motion/issues/3", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
961636297
(Small) Code Reestructuring Together with issue #28 , I'm trying to reorganise the code so that it becomes more clear. So datasets will become data (and I intend to, in the future, have a cleanup of that), and we need sub packages for fitting methods (as we will have 3: one taken from ngmix, our simple moments & @b-remy's model fitting). This is mostly renaming and reorganising (except for, again, #28, which I'm doing right now) but I'll then make a PR and ask @EiffL to go through it. You can open a draft PR, meaning it's not completely ready for review, but it can be convenient. I think this can be closed? right @andrevitorelli ? Yes, for now. I'm happy with the current structure (which didn't change much, tbh)
gharchive/issue
2021-08-05T09:37:58
2025-04-01T06:36:51.583701
{ "authors": [ "EiffL", "andrevitorelli" ], "repo": "CosmoStat/autometacal", "url": "https://github.com/CosmoStat/autometacal/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1330015275
Cleaner juno/app Overview Create app/upgrades dir to have all the upgrades in one place with versions and abstract out all keepers to app/keepers to make the the app/app.go cleaner and more maintainable. Problem Definition app/app.go is too big file. Proposal split app/app.go into app.go, upgrade/, keepers/ , module.go ref: osmosis/app provenance/app Hi there! I really liked your PR for the hub, this one: https://github.com/cosmos/gaia/pull/1580 I have been in seoul and unable to give it a proper review. Are the osmo and provenance styles different? Point blank: I agree with you. @faddat - refactor useful? might be some good practice to get warm @faddat provenance style is a little different in the sense they define there upgrades as a map with values as the upgrade function like this, then they have functions to deal with this map and create handlers out of this. In osmo this is more modular and we create a package itself per upgrade version, instead of a key value in map. I prefer the osmo way since it is cleaner and will be able to do alot more, includig handling forks better as well I'm very happy ot use the osmo style :D
gharchive/issue
2022-08-05T14:30:50
2025-04-01T06:36:51.588591
{ "authors": [ "Anmol1696", "faddat", "sascha1337" ], "repo": "CosmosContracts/juno", "url": "https://github.com/CosmosContracts/juno/issues/245", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1167699325
Front-page login for logged in accounts Subject: Front-page login for logged in accounts Description: Not necessarily a bug, but it threw me off. So clicking the couchers logo put me into the log-in page. I was expecting to land on almost any other page. I often use the site-logo as a sort of refresh beetween tasks. Results: Could it be possible to check to see if the user is logged in. If so, put them on the dashboard or some other site instead? Backend version: develop-bf8528d9 Frontend version: develop-bf8528d9 User Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/99.0.4844.51 Safari/537.36 Page: https://couchers.org/ User: myhgis (10850) Can't reproduce
gharchive/issue
2022-03-13T21:31:20
2025-04-01T06:36:51.609955
{ "authors": [ "CouchersBot", "lucaslcode" ], "repo": "Couchers-org/couchers", "url": "https://github.com/Couchers-org/couchers/issues/2680", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
995525147
⚠️ sandbox-api.piste.gouv.fr healthcheck has degraded performance In a467991, sandbox-api.piste.gouv.fr healthcheck (https://sandbox-api.piste.gouv.fr/minju/judilibre/v1.0/healthcheck) experienced degraded performance: HTTP code: 200 Response time: 2973 ms Resolved: sandbox-api.piste.gouv.fr healthcheck performance has improved in da7562e.
gharchive/issue
2021-09-14T02:47:52
2025-04-01T06:36:51.615808
{ "authors": [ "rhanka" ], "repo": "Cour-de-cassation/judilibre-uptime", "url": "https://github.com/Cour-de-cassation/judilibre-uptime/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2426271144
默认会在倒数第二个字符上显示背景色 IntelliJ IDEA 2024.1.4 (Community Edition) 插件版本:1.5.0.25-IDEA223 问题: 在未进行任何选择操作的情况下,倒数第二个字显示选择状态的背景色 如图最后一行‘s’看样子是选中状态,实际上我没有,光标在第六行正准备打字输入 禁用插件后状态会恢复正常 这是插件的诊断, 说你不是以新行结束,出于一些原因在idea上无法正常的标记在最末尾的位置 原来如此,按个回车果然没了
gharchive/issue
2024-07-23T23:19:46
2025-04-01T06:36:51.632803
{ "authors": [ "CppCXY", "imcloud" ], "repo": "CppCXY/EmmyLuaCodeStyle", "url": "https://github.com/CppCXY/EmmyLuaCodeStyle/issues/181", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
143087068
Testing out a new Mac OSX script I of course don't own a mac, (because PCMasterRace that's why) and I googled a bit finding the following below, can someone test this on there mac possibly to see if it works? Edit the old command text and change to this: bash -c 'while [ 0 ]; do date; node $(cd "$(dirname "$0")"; pwd)/lvlgen.js done' DOES NO ONE HAVE A MAC? lol Seems the agario community is intelligent enough to dont buy this crap :p Saw many hero members are on IOS/Mac, but there also not very intelligent lol I honestly think we should drop Mac development and leave it up for themselves to figure it out. That could be just me though :D I am with you @MastaCoder Mac and All Apple products are crap XD
gharchive/issue
2016-03-23T21:41:55
2025-04-01T06:36:51.634901
{ "authors": [ "MastaCoder", "christopherkardas", "hassanila97" ], "repo": "Cr4xy/agar-lvlgen", "url": "https://github.com/Cr4xy/agar-lvlgen/issues/31", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1212446911
Update credential documentation Summary and Scope Update to password change documentation to avoid out-of-order operation resulting in the inability to power components on and off. Issues and Related PRs Resolves CASMHMS-5472 for mainline Change will also be needed in release/1.2.5, release/1.2, release/1.0, release/0.9 Pull Request Checklist [x] Target branch correct /backport --dry-run release/1.2 release/1.1 release/1.0 /backport --dry-run release/1.2 release/1.1 release/1.0 /backport release/1.2
gharchive/pull-request
2022-04-22T14:38:15
2025-04-01T06:36:51.654455
{ "authors": [ "mjendrysik-hpe" ], "repo": "Cray-HPE/docs-csm", "url": "https://github.com/Cray-HPE/docs-csm/pull/1428", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2256269691
CRAYSAT-1839-sat-bootsys-waiter-subclass-addition Updating the sat bootsys kubelet start function to address chronjob creation failure IM:CRAYSAT-1839 Reviewer:Ryan Summary and Scope modified the do_kubelet_start function to add the wait after kubelet gets started. using the Waiter class that is already used in this module. Created a new Waiter subclass that implements the has_completed method to query the Kubernetes API to get the nodes in the Cluster. Once the query returns successfully, would be proceeding with the next step. Issues and Related PRs List and characterize relationship to Jira/Github issues and other pull requests. Be sure to list dependencies. Resolves CRAYSAT-1839 Testing Yet to be tested . Tested on: YTT Test description: will power off the nodes and power on the node to validate the functionality inclusion during the power on. Risks and Mitigations Minimal Pull Request Checklist [x] Version number(s) incremented, if applicable [x] Copyrights updated [x] License file intact [x] Target branch correct [x] CHANGELOG.md updated [ ] Testing is appropriate and complete, if applicable [ ] HPC Product Announcement prepared, if applicable Test output: ncn-m001:/mnt/shiva # sat bootsys boot --stage platform-services --ceph-timeout 120 The following Non-compute Nodes (NCNs) will be included in this operation: managers: ncn-m001 ncn-m002 ncn-m003 storage: ncn-s001 ncn-s002 ncn-s003 ncn-s004 workers: ncn-w001 ncn-w002 ncn-w003 ncn-w004 Are the above NCN groupings correct? [yes,no] yes INFO: Executing step: Ensure containerd is running and enabled on all Kubernetes NCNs. INFO: Executing step: Ensure etcd is running and enabled on all Kubernetes manager NCNs. INFO: Executing step: Start and enable kubelet on all Kubernetes NCNs. WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c987f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9ca7f70>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c87790>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8a9a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8abe0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9ca6a90>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8d250>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8eeb0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8ee20>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87adc0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87ad00>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87a0a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c9f040>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d400>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d190>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c99c40>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c99a90>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c99fd0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881be0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881220>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885250>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881130>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881d30>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881820>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c99ac0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c99c40>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aabe8850>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d4c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c9f6a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9cad1f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8853d0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885280>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885a00>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87deb0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d1f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d6a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c99fa0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c99fd0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aabe44c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885ca0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885610>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8857c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aaa731c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c767f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881af0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3ad1a3bb0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c99f70>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c99fd0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d1f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d4c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8d8e0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885640>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885580>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aaa73370>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885670>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8859a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885790>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d400>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87df10>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8a9a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3ad1a3b50>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aabe44f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881820>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aaa73490>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aaa73400>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aaa73910>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d2b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87deb0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885ca0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885f10>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885e50>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885ee0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87aca0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87aaf0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c9fb80>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c98490>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c988e0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c9f6a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885a60>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885340>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8856a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885b80>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885640>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885820>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d3a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881820>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c982b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9cae880>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c98940>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c982b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aab4fe80>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8854c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885520>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8852b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8853a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885df0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87ab50>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c9fbb0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c9f6a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885970>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8852b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885f10>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d3a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d3d0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d190>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c98070>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c98940>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9caefa0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8d430>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8d550>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9cae580>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881820>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881b80>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d1f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885220>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885790>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885640>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885be0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885580>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8d730>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885580>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8858b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885730>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885b50>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885910>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aab4f760>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9ca1a00>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9ca6dc0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c98070>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8d9d0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8d820>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8df40>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881070>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d430>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d400>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8d8b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885490>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8855b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885df0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885fa0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885610>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8aee0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aab4fe80>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885bb0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885b80>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8857c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8854c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d1f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8811c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881820>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c983d0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c98040>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8af70>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c98040>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c98940>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c8d8b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881d60>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d4c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d1f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa8854c0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885940>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885a30>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c87790>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c879d0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c87cd0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885b50>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa885d00>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa87d3a0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3aa881190>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9ca1fa0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9ca14f0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c98040>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c87580>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c87a00>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. WARNING:urllib3.connectionpool:Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c995b0>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c87f40>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:urllib3.connectionpool:Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError('<urllib3.connection.HTTPConnection object at 0x7fe3a9c87d30>: Failed to establish a new connection: [Errno 111] Connection refused')': /api/v1/nodes WARNING:root:MaxRetryError occurred while trying to connect to Kubernetes API. ERROR:/sat/venv/lib/python3.9/site-packages/sat/waiting.py:Waiting for condition "Kubernetes API availability check" timed out after 300 seconds INFO: Executing step: Recreate cron jobs that have become stuck INFO:sat.cli.bootsys.platform:Executing step: Recreate cron jobs that have become stuck DEBUG:csm_api_client.k8s:Couldn't load the in-cluster config: Service host/port is not set. (proceeding under the assumption that the config should be loaded from the kubeconfig file) the warning seems to be lot many, which may lead to confusion, even though its on the expected lines Updated output, post addressing the comments: ncn-m001:/mnt/shiva # sat bootsys boot --stage platform-services --ceph-timeout 120 The following Non-compute Nodes (NCNs) will be included in this operation: managers: ncn-m001 ncn-m002 ncn-m003 storage: ncn-s001 ncn-s002 ncn-s003 ncn-s004 workers: ncn-w001 ncn-w002 ncn-w003 ncn-w004 Are the above NCN groupings correct? [yes,no] yes INFO: Executing step: Ensure containerd is running and enabled on all Kubernetes NCNs. INFO: Executing step: Ensure etcd is running and enabled on all Kubernetes manager NCNs. INFO: Executing step: Start and enable kubelet on all Kubernetes NCNs. INFO: Waiting for Kubernetes API to become reachable... Waiting for condition "Kubernetes API available" timed out after 300 seconds INFO: Executing step: Recreate cron jobs that have become stuck WARNING: Jobs for cronjob "cray-dns-unbound-manager" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: Jobs for cronjob "hms-discovery" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. ncn-m001:/mnt/shiva # systemctl is-enabled kubelet enabled ncn-m001:/mnt/shiva # systemctl status kubelet ● kubelet.service - kubelet: The Kubernetes Node Agent Loaded: loaded (/usr/lib/systemd/system/kubelet.service; enabled; vendor preset: disabled) Drop-In: /usr/lib/systemd/system/kubelet.service.d └─10-kubeadm.conf /etc/systemd/system/kubelet.service.d └─10-kubelet.conf Active: active (running) since Thu 2024-04-25 06:46:32 UTC; 20min ago Docs: https://kubernetes.io/docs/ Main PID: 2977924 (kubelet) Tasks: 79 CGroup: /system.slice/kubelet.service └─ 2977924 /usr/bin/kubelet --bootstrap-kubeconfig=/etc/kubernetes/bootstrap-kubelet.conf --kubeconfig=/etc/kube> Apr 25 06:52:37 ncn-m001 kubelet[2977924]: E0425 06:52:37.079689 2977924 pod_workers.go:951] "Error syncing pod, skipping" er> Apr 25 06:52:51 ncn-m001 kubelet[2977924]: E0425 06:52:51.079839 2977924 pod_workers.go:951] "Error syncing pod, skipping" er> Apr 25 06:53:04 ncn-m001 kubelet[2977924]: E0425 06:53:04.080191 2977924 pod_workers.go:951] "Error syncing pod, skipping" er> Apr 25 06:53:17 ncn-m001 kubelet[2977924]: E0425 06:53:17.079126 2977924 pod_workers.go:951] "Error syncing pod, skipping" er> Apr 25 06:53:45 ncn-m001 kubelet[2977924]: I0425 06:53:45.778928 2977924 scope.go:110] "RemoveContainer" containerID="f58d126> Apr 25 06:53:45 ncn-m001 kubelet[2977924]: I0425 06:53:45.779894 2977924 scope.go:110] "RemoveContainer" containerID="f58d126> Apr 25 06:53:45 ncn-m001 kubelet[2977924]: E0425 06:53:45.781904 2977924 remote_runtime.go:296] "RemoveContainer from runtime> Apr 25 06:53:45 ncn-m001 kubelet[2977924]: E0425 06:53:45.781995 2977924 kuberuntime_container.go:798] failed to remove pod i> Apr 25 06:53:47 ncn-m001 kubelet[2977924]: I0425 06:53:47.796994 2977924 scope.go:110] "RemoveContainer" containerID="07bfd90> Apr 25 06:53:47 ncn-m001 kubelet[2977924]: I0425 06:53:47.797035 2977924 scope.go:110] "RemoveContainer" containerID="a90ee4d> ncn-m001:/mnt/shiva # kubectl cluster-info Kubernetes control plane is running at https://10.252.1.2:6442 CoreDNS is running at https://10.252.1.2:6442/api/v1/namespaces/kube-system/services/kube-dns:dns/proxy To further debug and diagnose cluster problems, use 'kubectl cluster-info dump'. @shivaprasad-metimath, from your testing, I can see that this code does not behave correctly. Specifically, see these lines: INFO: Waiting for Kubernetes API to become reachable... Waiting for condition "Kubernetes API available" timed out after 300 seconds INFO: Executing step: Recreate cron jobs that have become stuck WARNING: Jobs for cronjob "cray-dns-unbound-manager" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: Jobs for cronjob "hms-discovery" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. This shows that it timed out waiting for the Kubernetes API to become available. This means your has_completed method in your new KubernetesAPIWaiter class must never be returning True and allowing the waiter to complete. I looked into it, and I noticed that you are never loading the Kubernetes configuration, so it's not able to connect to the Kubernetes cluster. To fix this, please use the function load_kube_api from csm_api_client.k8s. This will properly ensure the Kubernetes configuration is loaded and then instantiate a CoreV1Api and return it to you. Note that when I tested using load_kube_api in a sat bash shell, I found another bug with calling list_node on the resulting CoreV1Api object. I addressed that in my pull request for CRAYSAT-1848 here: https://github.com/Cray-HPE/sat/pull/216 Please take a look at that. Once that's fixed in the main branch, you'll want to merge main into your feature/CRAYSAT-1740 branch and then rebase your CRAYSAT-1839-bootsys-waiter-subclass-addition branch on the feature branch to pull in the fix before you test again. While I was looking at other usages of the {{wait_for_completion}} method of various {{Waiters}}, I noticed that there is actually already a KubernetesAPIAvailableWatier in sat.cli.bootsys.k8s. It uses get_api_resources instead of list_node, and that's fine. All we're really trying to do is query the API, and that's probably the simplest API call you could make. As an added bonus, it doesn't look like get_api_resources has the same problem with the Kubernetes version mismatch as described in CRAYSAT-1848. The only thing missing from that class that you have in yours is an info log message the first time the waiter finds the Kubernetes API to be unreachable. I think it's fine to add such a log message. Please modify and use the existing KubernetesAPIAvailableWaiter instead of implementing another one. Latest output: ncn-m001:/mnt/shiva # sat bootsys boot --stage platform-services --ceph-timeout 120 The following Non-compute Nodes (NCNs) will be included in this operation: managers: ncn-m001 ncn-m002 ncn-m003 storage: ncn-s001 ncn-s002 ncn-s003 ncn-s004 workers: ncn-w001 ncn-w002 ncn-w003 ncn-w004 Are the above NCN groupings correct? [yes,no] yes INFO: Executing step: Ensure containerd is running and enabled on all Kubernetes NCNs. INFO: Executing step: Ensure etcd is running and enabled on all Kubernetes manager NCNs. INFO: Executing step: Start and enable kubelet on all Kubernetes NCNs. INFO: The Kubernetes API is currently unreachable. INFO: Executing step: Recreate cron jobs that have become stuck WARNING: Jobs for cronjob "cray-bos-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-bos-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '57bb6e09-ec2b-4dcb-a2bf-86e0a855d65b', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:15 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-bss-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-bss-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': 'be931aa2-c6ab-4f5d-9325-23c0d231e2c4', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:20 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-dns-unbound-manager" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-dns-unbound-manager" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '8c4d8af1-911d-4b90-a9ea-23881c3bef50', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:23 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-fas-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-fas-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '52dd4fe9-1419-4733-9527-0c6349b25ea5', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:26 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-hbtd-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-hbtd-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '61bc1a16-c0da-43b8-9386-4980c4f14642', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:29 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-hmnfd-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-hmnfd-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '2784415f-090b-4c01-acd9-89eed145a24b', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:29 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-power-control-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-power-control-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '6842288e-8752-4b75-ae21-902cbf0fb2f7', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:38 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-uas-mgr-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-uas-mgr-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '47650866-831a-4e66-bd36-6473a3f5f6cd', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:41 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "etcd-backup-pvc-snapshots-to-s3" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "etcd-backup-pvc-snapshots-to-s3" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '030149ce-48bf-4e2e-9103-84eede7bc86f', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:44 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "hms-discovery" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "hms-discovery" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '60fd74fb-e410-4c41-8fd8-46dc7748b734', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:47 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "sonar-sync" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "sonar-sync" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': 'ffd241fa-b628-4624-9d83-72816c5da21d', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '4ea23580-d341-47ea-a184-d76c9126b813', 'X-Kubernetes-Pf-Prioritylevel-Uid': '03d335b2-1a9f-4bed-979b-b187c7e83a43', 'Date': 'Tue, 30 Apr 2024 07:26:47 GMT', 'Content-Length': '545'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.17.72.31:443: connect: no route to host"}]},"code":500} WARNING: ncn-m001:/mnt/shiva # Waiting for condition "Kubernetes API available" timed out after 300 seconds ERROR: Fatal error in step "Start and enable kubelet on all Kubernetes NCNs." of platform services start: Failed to start kubelet: Kubernetes API not available Why is this error popped up even after I see kubelet is active and enabled Kubernetes api availability check seems to be working fine with expected output, during the chron job creation there is a delay observed leading to warnings. It could be observed only on this system. Complete log has been attached for reference. sat1.log 1st attempt: ncn-m001:/mnt/shiva # sat bootsys boot --stage platform-services --ceph-timeout 120 The following Non-compute Nodes (NCNs) will be included in this operation: managers: ncn-m001 ncn-m002 ncn-m003 storage: ncn-s001 ncn-s002 ncn-s003 ncn-s004 workers: ncn-w001 ncn-w002 ncn-w003 ncn-w004 Are the above NCN groupings correct? [yes,no] yes INFO: Executing step: Ensure containerd is running and enabled on all Kubernetes NCNs. INFO: Executing step: Ensure etcd is running and enabled on all Kubernetes manager NCNs. INFO: Executing step: Start and enable kubelet on all Kubernetes NCNs. INFO: Waiting up to 300 seconds for the Kubernetes API to become available INFO: The Kubernetes API is currently unreachable. INFO: Kubernetes API is available INFO: Executing step: Recreate cron jobs that have become stuck WARNING: Jobs for cronjob "cray-bos-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-bos-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': 'cc070e0a-fce8-45a1-8902-7e5d4ab8ff49', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:29 GMT', 'Content-Length': '493'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": context deadline exceeded","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": context deadline exceeded"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-bss-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-bss-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '1065b4ab-5747-4449-9ae6-cd000f659d59', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:32 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-dns-unbound-manager" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-dns-unbound-manager" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': 'a383ca98-5a4a-4bb0-962e-3ad778cac7a4', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:35 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-fas-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-fas-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': 'df5dc7f7-9a90-4f96-8de1-7daaf829166c', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:38 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-hbtd-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-hbtd-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '6eceeac2-152a-4ace-8c9b-d436c8f60c4f', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:41 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-hmnfd-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-hmnfd-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': 'da31639f-0f55-4b1e-8729-dc21aa9c4298', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:42 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-power-control-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-power-control-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '0a6b87f1-ed1b-461a-bb94-f535e0ec7738', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:45 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "cray-uas-mgr-bitnami-etcd-snapshotter" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "cray-uas-mgr-bitnami-etcd-snapshotter" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': 'c06ba9bd-cbbe-4e12-b24f-6be1ee537104', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:48 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "etcd-backup-pvc-snapshots-to-s3" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "etcd-backup-pvc-snapshots-to-s3" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': 'ef533cf7-2517-4304-bc21-7c78023f270c', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:51 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "hms-discovery" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "hms-discovery" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '1785c39e-98ed-4a44-88ae-d435d046ca78', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:54 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: WARNING: Jobs for cronjob "sonar-sync" in namespace "services" do not appear to be scheduled on time according to the cron job's schedule; recreating cron job. WARNING: An error occurred while re-creating cronjob "sonar-sync" in namespace "services": (500) WARNING: Reason: Internal Server Error WARNING: HTTP response headers: HTTPHeaderDict({'Audit-Id': '4cd734ff-ba4c-493c-bb7b-b6de10682b69', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Kubernetes-Pf-Flowschema-Uid': '206250a1-21b4-4b9d-9608-833e6a1583cd', 'X-Kubernetes-Pf-Prioritylevel-Uid': 'b5beaacd-765d-46f9-988c-cf2c75636da4', 'Date': 'Wed, 15 May 2024 10:23:57 GMT', 'Content-Length': '549'}) WARNING: HTTP response body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"Internal error occurred: failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host","reason":"InternalError","details":{"causes":[{"message":"failed calling webhook "validate.kyverno.svc-fail": Post "https://cray-kyverno-svc.kyverno.svc:443/validate/fail?timeout=10s\": dial tcp 10.23.152.167:443: connect: no route to host"}]},"code":500} WARNING: 2nd attempt: ncn-m001:/mnt/shiva # sat bootsys boot --stage platform-services --ceph-timeout 120 The following Non-compute Nodes (NCNs) will be included in this operation: managers: ncn-m001 ncn-m002 ncn-m003 storage: ncn-s001 ncn-s002 ncn-s003 ncn-s004 workers: ncn-w001 ncn-w002 ncn-w003 ncn-w004 Are the above NCN groupings correct? [yes,no] yes INFO: Executing step: Ensure containerd is running and enabled on all Kubernetes NCNs. INFO: Executing step: Ensure etcd is running and enabled on all Kubernetes manager NCNs. INFO: Executing step: Start and enable kubelet on all Kubernetes NCNs. INFO: Waiting up to 300 seconds for the Kubernetes API to become available INFO: Kubernetes API is available INFO: Executing step: Recreate cron jobs that have become stuck
gharchive/pull-request
2024-04-22T11:16:37
2025-04-01T06:36:51.906094
{ "authors": [ "annapoorna-s-alt", "haasken-hpe", "shivaprasad-metimath" ], "repo": "Cray-HPE/sat", "url": "https://github.com/Cray-HPE/sat/pull/214", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1948767020
Translations update from Hosted Weblate Translations update from Hosted Weblate for Metadata Remover/Metadata. It also includes following components: Metadata Remover/Metadata Remover Current translation status: Thank you very much for your support! 😘
gharchive/pull-request
2023-10-18T04:08:45
2025-04-01T06:36:51.915206
{ "authors": [ "CrazyMarvin", "weblate" ], "repo": "Crazy-Marvin/MetadataRemover", "url": "https://github.com/Crazy-Marvin/MetadataRemover/pull/44", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1953904403
🛑 sneak.page is down In 4b6a6a9, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in 4cc7895 after 8 minutes.
gharchive/issue
2023-10-20T09:34:22
2025-04-01T06:36:51.917864
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/325", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1957740294
🛑 sneak.page is down In 52a693a, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in 83127fe after 11 minutes.
gharchive/issue
2023-10-23T18:13:42
2025-04-01T06:36:51.920562
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/389", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2342105386
🛑 sneak.page is down In 7169431, sneak.page (https://sneak.page) was down: HTTP code: 404 Response time: 107 ms Resolved: sneak.page is back up in ed322a0 after 8 minutes.
gharchive/issue
2024-06-09T08:40:10
2025-04-01T06:36:51.923237
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/4325", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1968888242
🛑 sneak.page is down In 40d327f, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in 6e9a863 after 16 minutes.
gharchive/issue
2023-10-30T17:33:03
2025-04-01T06:36:51.925948
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/516", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2466951547
🛑 sneak.page is down In 87763e7, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in 20d83c2 after 8 minutes.
gharchive/issue
2024-08-14T22:37:55
2025-04-01T06:36:51.929485
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/5322", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1977197542
🛑 sneak.page is down In 0e4fd12, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in 36fa221 after 16 minutes.
gharchive/issue
2023-11-04T06:30:48
2025-04-01T06:36:51.931945
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/590", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
266051597
Store in DB Hello! Thank you for this package, it is very usefull. In my project I want to do this. "User One" logges in, put several items to the cart and loggs out. After a while he logges in again and sees items in his cart. I know than I may use store method, but I don't understand how, cause I have a "CartAlreadyStoredException". Where may I put code for this method? My code for adding products: public function addProductToCart(Request $request, $id) { if ($request->has('amount')){ $amount = $request->amount; } else { $amount = 1; } $product = Product::find($id); Cart::instance('shopping')->add($id, $product->name, $amount, $product->price); if(Auth::check()){ Cart::instance('shopping')->store(Auth::user()->id); //error! } return back(); } and for displaying: (now it shows items that are in the session, but I want to swith between cart in the session for not auth users and between cart in the DB for auth users) public function cart() { $cartItems = Cart::instance('shopping')->content(); return view ('orders.cart', compact('cartItems')); } To sum up, I need to use session when user is guest in and DB when user is logged in. Once again, thanks for package. I hope someone can help me:) I'm afraid you're using the 'store' method in another way than it's designed for. The idea is that you can 'store' the cart for a later point in time, and 'restore' it when the user wants it again. It's not really designed to store the cart in the database 'realtime' so to say. Thanks for your answer) If someone has this problem, that's my suggestion. In logout method: public function logout(Request $request) { Cart::instance('shopping')->restore(Auth::user()->id); Cart::instance('shopping')->store(Auth::user()->id); $this->guard()->logout(); $request->session()->invalidate(); return redirect(url()->previous()); } @humamalamin If I am logged in I store data in the session. My updated logout function public function logout(Request $request) { //delete old cartitems DB::table('shoppingcart')->where([ ['identifier', Auth::user()->id], ['instance', 'shopping'] ])->delete(); DB::table('shoppingcart')->where([ ['identifier', Auth::user()->id], ['instance', 'wishlist'] ])->delete(); //save new cart items Cart::instance('shopping')->store(Auth::user()->id); Cart::instance('wishlist')->store(Auth::user()->id); $this->guard()->logout(); $request->session()->invalidate(); //clear session data. cart became empty return redirect(something_url); } When I'm logged in I combine data from DB (if it is exists) and data from the cart (if customer add smth already) public function login(Request $request) { /* * ..... * checking validation and logging user in. if OK */ // get cart from DB if it exists $storedCartItems = DB::table('shoppingcart')->where([ ['identifier', Auth::user()->id], ['instance', 'shopping'] ])->value('content'); // get wishlist from DB if it exists $storedWishlistItems = DB::table('shoppingcart')->where([ ['identifier', Auth::user()->id], ['instance', 'wishlist'] ])->value('content'); $storedCartItems = \unserialize($storedCartItems); $storedWishlistItems = \unserialize($storedWishlistItems); // check if count of each product in the store is more than in the cart and more than 0 (only for cart) if($storedCartItems){ foreach ($storedCartItems as $item){ Cart::instance('shopping')->add($item->id, $item->name, $item->qty, $item->price)->associate('App\Product'); // if it passes, I'll add them to the cart in the session if (($item->model->qty > 0) & ($item->model->qty < $item->qty)){ Cart::instance('shopping')->update($item->rowId, $item->model->qty); // if it does not pass, I will not add them to the cart in the session } elseif ($item->model->qty == 0){ Cart::instance('shopping')->remove($item->rowId); } } } //add items from wishlist from DB to the wishlist items in the session if($storedWishlistItems){ foreach ($storedWishlistItems as $item){ Cart::instance('wishlist')->add($item->id, $item->name, $item->qty, $item->price)->associate('App\Product'); } } // return redirect or smth else } That works for me. Hope it can help you)
gharchive/issue
2017-10-17T09:21:44
2025-04-01T06:36:51.974369
{ "authors": [ "Crinsane", "ellgibug" ], "repo": "Crinsane/LaravelShoppingcart", "url": "https://github.com/Crinsane/LaravelShoppingcart/issues/397", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
310786658
Share dialog Hi, how can I show the share dialog before posting on Facebook? You can create your own ui for sharing. We made it flexible so that you can create your own UX, instead of using fb default dialog.
gharchive/issue
2018-04-03T10:50:15
2025-04-01T06:36:51.990041
{ "authors": [ "autosoftmultimedia", "rdelrosario" ], "repo": "CrossGeeks/FacebookClientPlugin", "url": "https://github.com/CrossGeeks/FacebookClientPlugin/issues/16", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1484947552
[C-288] Create tooltips for severals terms in web app Several users asked for explanations/defintions for different terms, including "Active" members Identities Engagement level Reach Attributes Solution: Create tooltips for those terms in the web app From SyncLinear.com | C-288 @jonathimer @joanagmaia Is this issue still valid? If so, please can you provide more context. @nu Pinging @joanagmaia @nunoeufrasio again :) Is this issue still valid? If so, please can you provide more context. duplicate
gharchive/issue
2022-12-08T16:06:46
2025-04-01T06:36:52.001797
{ "authors": [ "jonathimer", "peoray" ], "repo": "CrowdDotDev/crowd.dev", "url": "https://github.com/CrowdDotDev/crowd.dev/issues/328", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1665287169
Bug Fix // Corrected Typo Changes proposed ✍️ Corrected typos — "coulnd't" ➡️ "couldn't" Added period to end of full sentences Screenshots (front-end changes only) - N/A Checklist ✅ [X] Label appropriately with Feature, Enhancement, or Bug. [N/A] Tests are passing [N/A] New backend functionality has been unit-tested. [N/A] Environment variables have been updated: [N/A] Local frontend configuration: frontend/.env.dist.local, frontend/.env.dist.composed. [N/A] Local backend: backend/.env.dist.local, backend/.env.dist.composed. [N/A] Configuration docs have been updated. [N/A] Team members only: update environment variables in override, staging and production env. files and trigger update config script. [N/A] API documentation has been updated (if necessary) (see docs on API documentation). [X] Quality standards are met. [ ] All changes have been tested in a staging site. [ ] All changes are working locally running crowd.dev's Docker local environment. Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
gharchive/pull-request
2023-04-12T21:27:41
2025-04-01T06:36:52.008944
{ "authors": [ "CLAassistant", "erinmikailstaples" ], "repo": "CrowdDotDev/crowd.dev", "url": "https://github.com/CrowdDotDev/crowd.dev/pull/753", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
674306245
code plz? Could you provide the code of GroupIM, plz? Thx! Thanks for your interest in our work, the code will be released by the end of the month. Thanks for your interest in our work, the code will be released by the end of the month. Tomorrow is the end of this month😃 Thanks for the reminder, the code is up now!
gharchive/issue
2020-08-06T13:16:34
2025-04-01T06:36:52.010731
{ "authors": [ "Jhy1993", "aravindsankar28", "zoeleesss" ], "repo": "CrowdDynamicsLab/GroupIM", "url": "https://github.com/CrowdDynamicsLab/GroupIM/issues/1", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1827647199
Fix Out of range errors in Data+String.swift Hi I love the library and use it in my podcast app HyperCatcher! I ran into what looks like an out of range error on this mp3 file: https://api.substack.com/feed/podcast/135258128/8ebf19eff23e3b34077912c5ff4f40b5.mp3 From this podcast feed: https://www.latent.space/feed This PR is the fix I implemented in my app Ah - nice catch. I'm not in a position to test this just at the minute. Just looking over it - the .single changes look good, but I'm not sure what the changes inside the .double case achieve exactly? Is it changing it to look forward instead of backwards? Ah - nice catch. I'm not in a position to test this just at the minute. Just looking over it - the .single changes look good, but I'm not sure what the changes inside the .double case achieve exactly? Is it changing it to look forward instead of backwards? Sorry for the delay, I was at a hackathon last week when I worked on this. I did add the .double case changes in my app, but looking back at this I don't remember why so I removed them for now. The .single case was where I was actually seeing the crash so the .double case is probably fine as it is anyway. Thanks for this!
gharchive/pull-request
2023-07-29T20:29:27
2025-04-01T06:36:52.030466
{ "authors": [ "HendX", "normand1" ], "repo": "CrunchyBagel/OutcastID3", "url": "https://github.com/CrunchyBagel/OutcastID3/pull/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1160697235
Add CryanCodes File created a branch, created a file and made a commit, and opened a pull request created a branch, created a file and made a commit, and opened a pull request
gharchive/pull-request
2022-03-06T19:40:56
2025-04-01T06:36:52.047753
{ "authors": [ "CryanCode" ], "repo": "CryanCode/github-slideshow", "url": "https://github.com/CryanCode/github-slideshow/pull/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
990997177
Add heroku deploy support Can you add heroku deploy support? I've never deployed something to Heroku before and it's not really something that I'm interested in figuring out. What would the benefit be? I've never deployed something to Heroku before and it's not really something that I'm interested in figuring out. What would the benefit be? heroku offers free hours for 20days and if we add CC it allows another 15days so it means we can host our app for free and unlimited bandwidth. Its easy and offers high speed bandwidth The server that https://av-converter.com runs on has unlimited bandwidth and as I am paying for the server, it is free for everyone else including you. Adding Heroku support to this project is not something that I want to spend my time on figuring out, at least not at this moment in time. You are free to give it a go yourself and submit a pull request if you are successful.
gharchive/issue
2021-09-08T10:45:25
2025-04-01T06:36:52.050263
{ "authors": [ "CrypticSignal", "askfriends" ], "repo": "CrypticSignal/av-converter", "url": "https://github.com/CrypticSignal/av-converter/issues/21", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
625906945
Add support for DMM The interface is very similar to compound. All mTokens have the same number of decimals as their underlying counterpart. The exchangeRate is always a number with 18 decimal places, regardless of the token. Master Controller - 0x4CB120Dd1D33C9A3De8Bc15620C7Cd43418d77E2 You can call #getDmmTokenIds to get all of the tokens by their ID and iterate through them, if you want. There are the respective functions in there for getting the underlying token by address or querying on the mToken address if you only have the underlying (eg DAI --> mDAI) underlyingTokenAddressToDmmTokenIdMap mDAI -0x06301057D77D54B6e14c7FafFB11Ffc7Cab4eaa7 mETH - 0xdF9307DFf0a1B57660F60f9457D32027a55ca0B2 mUSDC - 0x3564ad35b9E95340E5Ace2D6251dbfC76098669B Minting is as simple as calling mint(uint underlyingAmount) where underlyingAmount is the amount of underlying you want to send into the contract. Note, a token approval is needed for calling mint on the underlying contract, where the spender is set to the mToken contract address. Redeeming mTokens for underlying, plus interest, is done through redeem(uint amount) where amount is the amount of mTokens to be sent to the contract and redeemed. Note, no token approvals are required to redeem mTokens. Thanks!
gharchive/issue
2020-05-27T17:54:45
2025-04-01T06:36:52.057369
{ "authors": [ "coreycaplan3", "k06a" ], "repo": "CryptoManiacsZone/1inchProtocol", "url": "https://github.com/CryptoManiacsZone/1inchProtocol/issues/21", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2581874777
Integrate Google Drive/Dropbox for Audiobook Uploads Description: Allow users to upload their own audiobooks from cloud services like Google Drive or Dropbox. This feature will enable users to sync their own audiobook collections with LisBook. Acceptance Criteria: Add "Upload from Google Drive" and "Upload from Dropbox" options in the app. Ensure proper handling of file formats and uploads. Test uploading from both cloud services. I'd like to give it a shot. Please assign it to me. Sure @dexterousdhruv I would love to see some cool contribution from your side. 😄. Don't Forget to ⭐ our repo and happy coding 👨🏼‍💻 @dexterousdhruv are you working on it ? If you need any help let me know I'm on it. Great @dexterousdhruv go ahead @Ctoic My question is where would we store the files, once the user uploads it from his respective cloud provider ? Guess what, we have to set up an express server. What do you say ?
gharchive/issue
2024-10-11T17:35:07
2025-04-01T06:36:52.060548
{ "authors": [ "Ctoic", "dexterousdhruv" ], "repo": "Ctoic/Lisbook", "url": "https://github.com/Ctoic/Lisbook/issues/35", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2619364571
Memory Usage Misreported on ARK: SA Operating System Unknown AMP Version and Build Date 2.6 AMP Release Stream Mainline I confirm that [X] I have searched for an existing bug report for this issue. [X] I am using the latest available version of AMP. [X] my operating system is up-to-date. Intended Action N/A Expected Behaviour N/A Actual Behaviour N/A Reproduction The memory usage is misreporting after the 2.6 upgrade. This is resolved in the release coming tomorrow.
gharchive/issue
2024-10-28T19:35:34
2025-04-01T06:36:52.064781
{ "authors": [ "IceOfWraith" ], "repo": "CubeCoders/AMP", "url": "https://github.com/CubeCoders/AMP/issues/1199", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
473672795
BUG: No console output using Minecraft module after upgrade to 1.8.3.0 Bug Report System Information Operating System: Ubuntu Server 18.04/Kernel 4.15.0-55-generic AMP version and build date: v1.8.3.0 built 25/07/2019 17.10 Which AMP release stream you're using: Mainline I confirm: [x] that I have searched for an existing bug report for this issue. [x] that I am using the latest available version of AMP. [x] that my operating system is up-to-date. Symptoms After upgrading my instances to 1.8.3.0, console output is not working in the Minecraft instance web interface. The game is accpeting commands from the server but no text is showing on the console tab in the web interface Reproduction Upgrade an existing 1.8.2.2 instance to 1.8.0.3 and check the console tab after starting the server (i have not tried creating a new instance) After clearing my browser cache, the console tab started working again. Please close this bug out as resolved. Thanks! After clearing my browser cache, the console tab started working again. Please close this bug out as resolved. Thanks! You have a button at the bottom of the page that lets you "close" the issue.
gharchive/issue
2019-07-27T20:03:38
2025-04-01T06:36:52.069636
{ "authors": [ "omega9380", "pizzafox" ], "repo": "CubeCoders/AMP", "url": "https://github.com/CubeCoders/AMP/issues/131", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1792794576
General adjustments in Taurus and Beowulf 🤔 What type of PR is this? (check all applicable) [ ] 🍕 Addition [ ] ⌨️ Productivity [X] 🐛 Bug Fix [ ] 🔥 Optimization [ ] ⚙️ Configuration [ ] 🌟 Quality Of Life [X] ✨ Enhancement [ ] 📝 Documentation 📝 Description fix and renew some more things, maybe in the next one I will fix the accessories of the taurus revolver and the m249 maybe it will fix it 🖼️ Screenshots/Recordings Bug: Looks good to me. Awaiting review from @Desoroxxx. I already added the changes to the changelog
gharchive/pull-request
2023-07-07T05:16:26
2025-04-01T06:36:52.073233
{ "authors": [ "Braiam23", "Paneedah" ], "repo": "Cubed-Development/Modern-Warfare-Cubed", "url": "https://github.com/Cubed-Development/Modern-Warfare-Cubed/pull/189", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1580788036
data-magnetic Hi again, I added this tag (data-magnetic), but it didn't work. I think it's disabled! Thank you. For data-magnetic you have to wrap your element with an div that holds data-cursor-stick. Heres an exaple: <div data-cursor-stick='#stick-here'> <div id='stick-here'> <BiMenu /> </div> </div> @sagarkhadka thank you for your time. Yes, I know this point. I meant the magnet like the picture below @HasanSibakhi For the effect I think cuberto has given different repo. You can find it here. Maybe this will help.
gharchive/issue
2023-02-11T10:56:51
2025-04-01T06:36:52.076782
{ "authors": [ "HasanSibakhi", "sagarkhadka" ], "repo": "Cuberto/mouse-follower", "url": "https://github.com/Cuberto/mouse-follower/issues/17", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2020701233
🛑 lemmy.best is down In 52131d7, lemmy.best (https://lemmy.best) was down: HTTP code: 0 Response time: 0 ms Resolved: lemmy.best is back up in 37cc728 after 4 minutes.
gharchive/issue
2023-12-01T11:34:45
2025-04-01T06:36:52.079218
{ "authors": [ "Curly-Mo" ], "repo": "Curly-Mo/lemmy.best-upptime", "url": "https://github.com/Curly-Mo/lemmy.best-upptime/issues/217", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2453898281
🛑 lemmy.best is down In 2ec437d, lemmy.best (https://lemmy.best) was down: HTTP code: 0 Response time: 0 ms Resolved: lemmy.best is back up in f9fbe1f after 7 minutes.
gharchive/issue
2024-08-07T16:38:19
2025-04-01T06:36:52.081580
{ "authors": [ "Curly-Mo" ], "repo": "Curly-Mo/lemmy.best-upptime", "url": "https://github.com/Curly-Mo/lemmy.best-upptime/issues/729", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1511500606
🛑 Cust+ is down In 6d50152, Cust+ (https://cust.cc) was down: HTTP code: 0 Response time: 0 ms Resolved: Cust+ is back up in ab08706.
gharchive/issue
2022-12-27T07:00:36
2025-04-01T06:36:52.084593
{ "authors": [ "xtyxtyx" ], "repo": "CustedNG/upptime", "url": "https://github.com/CustedNG/upptime/issues/278", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2080924890
🛑 必应一(镜像)bingo.0x.no is down In 9ad9621, 必应一(镜像)bingo.0x.no (https://bingo.0x.no/) was down: HTTP code: 0 Response time: 0 ms Resolved: 必应一(镜像)bingo.0x.no is back up in 0603871 after 8 minutes.
gharchive/issue
2024-01-14T22:31:27
2025-04-01T06:36:52.092748
{ "authors": [ "Cutiei" ], "repo": "Cutiei/Uptime", "url": "https://github.com/Cutiei/Uptime/issues/1322", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2087118000
🛑 搜索一So.h-e.top is down In fd268cf, 搜索一So.h-e.top (https://so.h-e.top/) was down: HTTP code: 521 Response time: 298 ms Resolved: 搜索一So.h-e.top is back up in bd4d107 after 10 minutes.
gharchive/issue
2024-01-17T22:30:20
2025-04-01T06:36:52.095485
{ "authors": [ "Cutiei" ], "repo": "Cutiei/Uptime", "url": "https://github.com/Cutiei/Uptime/issues/1411", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2358525163
#561: Add component and services for tools CycloneDX spec 1.5 deprecated an array of tools in bom.metadata and instead prefers object with an array of components and an array of services. This PR implements that. This works de-serializing a Syft SBOM with a tool section like so: "metadata": { "timestamp": "2024-06-10T13:06:52-08:00", "tools": { "components": [ { "type": "application", "author": "anchore", "name": "syft", "version": "1.4.1" } ] }, "component": { "bom-ref": "08329a07b4eb8eac", "type": "file", "name": "./" } }, Next up: docs, XML (de)serialization code, and tests. fixes #561 a feature i would love to see in the end: metadata tools converters, that allow me to have a bunch of Components and Services , and when normalizing to CycloneDX 1.3, the Compoennts and Servioces are converted to Tools in the resulting XML/JSON. So that i dont loose any data... i dislike the current concept of ToolRepository that holds tools, services and components at the same time. This is an abomination of a data type, and teaching people how to use it properly will just be a hack of an effort. i'd rather have a type ToolRepository that holds services and components and no Tool! So, Metadata.tools would be truly Union[ToolRepository, SortedSet[Tool]]. The (de)serialize would be handled by a (de)normaliser (aka "Helper" and thats it. Is there something that speaks against this very simple and still pythonic solution? @jkowalleck: So drivers for my approach were two-fold: I am not 100% acquainted with the use cases of the the library; that is how people use the library. I was trying to maintain 100% backward compatibility. Thus, the reason for ToolRepository which combined the three types. I wanted to be able to initialize and empty BomMetaData (which is allowed now) and then do: # old code still works bom.metadata.tools.add(....) # or bom.metadata.tools = my_sorted_set_of_tools as well as be able to do: # new code would work as expected bom.metadata.tools.components.add(...) bom.metadata.tools.services.add(...) If we don't have an object in bom.metadata.tools which responds to components and services attributes, then calling those attributes will generate an exception, and cause unexpected behavior. At least in my mind, I would expect to be able to instantiate a Bom, and then call bom.metadata.tools.components.add(...) or bom.metadata.tools.components = ... a feature i would love to see in the end: metadata tools converters, that allow me to have a bunch of Components and Services , and when normalizing to CycloneDX 1.3, the Compoennts and Servioces are converted to Tools in the resulting XML/JSON. So that i dont loose any data... That would be really neat. Do you mean a stand-alone tool? Or built in to the new type? Is there a way to know in the normalizing functions which version we're serializing for? I did not see that anywhere in the docs, but it would be great if we could do that. re https://github.com/CycloneDX/cyclonedx-python-lib/pull/635#issuecomment-2177181574 Thank you for the explanation. I understand the reasoning behind the current implementation of ToolsRepo now. For sake of usability, you are right to using it. A need for API backwards compatibility is not really needed. Clean code and documentation are more important. An improvement I could imagine: have it a much simpler, documented container-object like so: class ToolsRepository: """our implementation of the tools repo""" tools: SortedSet[Tool] """DEPRECATED tools""" components: SortedSet[Component] """docstring here..." services: SortedSet[Service] """docstring here..." def __init__(self, *, components: Optional[Iterable[Components]] = None, services: OptionalIterable[Service]] = None, # Deprecated in v1.x tools: OptionalIterable[Tool]] = None, ): if tools: warn("deprecation message here...", DeprecationWarning) self.tools = SortedSet(tools or ()) self.components = SortedSet(components or ()) self.services = SortedSet(services or ()) Thanks for the explanation. Given the comment here: https://github.com/CycloneDX/cyclonedx-python-lib/blob/main/cyclonedx/model/init.py#L1079 I'd like to try to keep backward compatibility for now. The code to maintain this is probably only 60 to 70 or so lines more than breaking backward compatibility, and once the support for List[Tool] goes away completely, the semantics for tools.components and tools.services won't change. Everything seems to be working, save the problem I mentioned here: https://cyclonedx.slack.com/archives/CVA0QJEVA/p1718821402381719 Tests for new functionality are in place. Pretty sure I didn't do it the best way, but just wanted to show the new functionality works. :) We can re-work the tests before merge if need be. BTW, maintains 93% test coverage. I have a few more code paths I want to test as well. @jkugler, before we can use your contribution, we need you to sign-off your commits. here is why this is required and what it implies: https://github.com/CycloneDX/cyclonedx-python-lib/blob/main/CONTRIBUTING.md#sign-off-your-commits here is a step-by-step instruction on how to do this: https://github.com/CycloneDX/cyclonedx-python-lib/pull/635/checks?check_run_id=26536713378 There is a decision we'll need to make in how to handle the current behavior in BomMetaData: if not tools: self.tools.add(ThisTool) This adds the CycloneDX information if none is provided when initializing BomMetaData, and the will generate an error if someone, later does this: bom.metadata.tools.components = ... # or bom.metadata.tools.services = However, there will not be an error if a developer does bom.metadata.tools.components.add(...) or similar with services, and if there are components or services, those will be rendered and not the [Tools]. If we're OK with that behavior, that is, telling users to provide a ToolsRepository themselves, or use .add(), I'm OK with that. @jkugler, before we can use your contribution, we need you to sign-off your commits. Yes, I did that in https://github.com/CycloneDX/cyclonedx-python-lib/pull/635/commits/2bbd659eec2ac6711da14e6f265c208b4ee61ccb I didn't realize I would need to do that for every commit, as I assumed the PR would be squash merged into a single commit, and thus include the sign-off. I'll get it fixed. re: https://github.com/CycloneDX/cyclonedx-python-lib/pull/635#issuecomment-2183494986 all the effort to make the thing backwards compatible? Well, the old Metadata.tools was of type SortedSet[Tool]. The new one implements a subset of SortedSet's functionality. If you really wanted to make it backwards compatible, then ToolsRepository MUST extend SortedSet. Then most of your concerns would be solved, right? re: https://github.com/CycloneDX/cyclonedx-python-lib/pull/635#issuecomment-2183523501 it will be squashed on merge, that is true, but only if each and every commit was signed-off. Otherwise, we would quash/merge a thing that never was legally usable :-) re: https://github.com/CycloneDX/cyclonedx-python-lib/pull/635#issuecomment-2183605697 I could inherit from SortedSet. I'm not sure if that's necessary, though. I guess by "backward compatible," I was more shooting for "all the existing tests pass" as that should indicate all existing (supported) use cases continue to work. Since we don't even document the BomMetaData class (nor BomMetaData.tools), I think it is a reasonable expectation that "tools" Set. I took that in to account with these functions in ToolsRepository: def __len__(self) -> int: return len(self._tools) def __bool__(self) -> bool: return any([self._tools, self._components, self._services def __getattr__(self, name: str) -> Any: """ Enables us to behave as list of tools to maintain backward compatibility. Returns: An attribute of SortedSet """ return getattr(self._tools, name) def __iter__(self) -> Iterator[Tool]: """ Also part of acting as a list of tools Returns Iterator[Tool] """ for t in self._tools: yield t That will keep the behavior of the set, passing all unknown attributes to SortedSet[Tools], and thus exposing the entire public interface of SortedSet[Tools]. I think we'll be good. If we want to release a breaking change, and call this out, I'm not against that, but I think most of the code out there using this library will continue to work. I know mine will. :) @jkugler please add test fixtures/models with the new features you've added? they go to https://github.com/CycloneDX/cyclonedx-python-lib/blob/main/tests/_data/models.py the new test fixture should be built by functions called like get_bom_<something>(). see also: https://github.com/CycloneDX/cyclonedx-python-lib/blob/49a93a03b38574f264a49e9515cd6aa7b0b0f4c5/tests/_data/models.py#L1128-L1165 @jkugler re https://github.com/CycloneDX/cyclonedx-python-lib/pull/635#issuecomment-2187510084 well, changing a property BomMetaData.tools from SortedSet[Tool] to something different is considered a breaking change. Why bother having backwards compatibility when we have a breaking change anyway? Since we don't even document the BomMetaData class (nor BomMetaData.tools) We do. A type annotation is considered documentation. it is even rendered as such, see https://cyclonedx-python-library.readthedocs.io/en/latest/autoapi/cyclonedx/model/bom/index.html#cyclonedx.model.bom.BomMetaData.tools well, changing a property BomMetaData.tools from SortedSet[Tool] to something different is considered a breaking change. Why bother having backwards compatibility when we have a breaking change anyway? I see what you mean. I was thinking about "breaking behavior" in this case. We can publish a breaking change, but I hope the way I've done things will results in no needed code changes for most of our users. We do. A type annotation is considered documentation. it is even rendered as such, see https://cyclonedx-python-library.readthedocs.io/en/latest/autoapi/cyclonedx/model/bom/index.html#cyclonedx.model.bom.BomMetaData.tools Somehow, I missed that. I apologize. missing test cases: craft a BOM with metadata.tools having 1 Component, 1 Service, and 1 Tool - render BOM to XML/JSON 1.0 to 1.6 without loosing essential data and schema-validate the result this should result in a set of 3 Tool in JSON/XML - Component/Service were converted to Tool, as all must be Tool - to be schema compatible craft a BOM with metadata.tools having 1 Component, 1 Service, and NO Tool - render BOM to XML/JSON 1.0 to 1.6 without loosing essential data and schema-validate the result this should show how in JSON/XML 1.4 and before: this should result in a set of 2 Tool - Component/Service were converted to Tool - to be schema compatible in JSON/XML 1.5 and later: Component/Service were NOT converted to Tool re: https://github.com/CycloneDX/cyclonedx-python-lib/pull/635#issuecomment-2190086101 There are many fields in Component and Service which do not map to the fields available in Tool. Silently converting those objects to Tool would quietly discard a host of information and the user would be none the wiser, and would assume all the information they have added has been rendered to the SBOM. In this case, there are five fields in Tool, 27 fields in Component, and 17 in Service. This would actually cause a good deal of data loss while the user assumes the SBOM will contain all data they have added. I would propose one of the two possibilities: On the "quiet" side: print out a large warning if there is information in SortedSet[Component] or SortedSet[Service] and we are rendering in CDX < 1.5. On the "loud" side: refuse to render CDX < 1.5 when there is information in SortedSet[Component] or SortedSet[Service] and raise some kind of exception. To minimize data loss, and to subscribe to the "principle of least surprise," I would direct a user who wanted to produce a CDX <= 1.4 SBOM. to use Tool, and CDX >= 1.5 to use component/service. That said, if we do the auto-conversion, where and how will we document this so it is obvious to users of the library? re https://github.com/CycloneDX/cyclonedx-python-lib/pull/635#issuecomment-2190165893 i'd vote for option 3): silently transform and accept loss of unavailable data, while carrying over all available data. On rendering, we also do not print a warning if a property is populated that is not available in the target version. This library adheres to the CycloneDX specification. If a user knowingly uses data models that are not available in the target version, they would not be surprised to lose this data. :-) The target audience is well-informed software-developed. The model transform on rendering to lower versions is a convenience feature may people need. We would put it in the library, so others don't need to reinvent the wheel, and since we already adhere to the spec and are the experts in the field. @jkugler, Could you rebase on latest master and fix the conflicts. I had to do some style-changes(#643) to make the review easier. Sorry for the inconvenience. @jkowalleck Is this what you had in mind? https://github.com/CycloneDX/cyclonedx-python-lib/pull/635/commits/1a639a678d35c17fd7dab660348f7952599ed10d @jkowalleck For the "auto conversion" of Components and Services to Tools for older schemas, does this mapping work? Tools: Vendor Name Version Hashes externalReferences Components: Supplier - > Vendor? Or Author -> Vendor? Or Publisher -> Vendor? Name -> Name Version -> Version Hashes -> Hashes externalReferences -> externalReferences Services Provider -> Vendor? Name -> Name Version -> Version No Hashes externalReferences -> externalReferences I'm still really concerned about the amount of data that would be lost, but we can go this route if you think it best. @jkowalleck Is this what you had in mind? 1a639a6 yes. that looks great. @jkowalleck For the "auto conversion" of Components and Services to Tools for older schemas, does this mapping work? Tools: * Vendor * Name * Version * Hashes * externalReferences Components: * Supplier - > Vendor? Or Author -> Vendor? Or Publisher -> Vendor? Group -> Vendor Author, Publisher etc are missing in Tool - that is why Component was introduced as a possible item-class of the tools repository. * Name -> Name * Version -> Version * Hashes -> Hashes * externalReferences -> externalReferences Services * Provider -> Vendor? Provider -> Vendor * Name -> Name * Version -> Version * No Hashes * externalReferences -> externalReferences I'm still really concerned about the amount of data that would be lost, but we can go this route if you think it best. THink of it more like this: downstream users might loose some data, but they know about this - but, most importantly, we save as much data as possible. Code was added to "down-convert" from new Tools to old Tools. Some snapshot updates. Will look more and see what additional testing needs to be done. Also need to update XML rendering. @jkowalleck: craft a BOM with metadata.tools having 1 Component, 1 Service, and 1 Tool - render BOM to XML/JSON 1.0 to 1.6 without loosing essential data and schema-validate the result this should result in a set of 3 Tool in JSON/XML - Component/Service were converted to Tool, as all must be Tool - to be schema compatible As currently coded, this cannot be done, because an exception will be thrown if one attempts to have a Tool and a Comonent /Service. Is that acceptable? Or are you proposing we accept all conflicting objects, and down-convert if there are Components or Services with Tool? The spec says Tools XOR Components/Services. I would think if someone was going to use Components/Services, they would want that detail and would want to be warned they are still using Tool and need to upgrade/convert that to a Component. Also, if they do not add Components/Services in BomMetadata upon initialization, and it gets a default ThisTool, then they would never be warned when they add Components/Services and it's down-converted. I would rather be explicit about "you shouldn't do that" instead of them seeing the output and wondering why their Components/Services were converted to SBOMs. craft a BOM with metadata.tools having 1 Component, 1 Service, and NO Tool - render BOM to XML/JSON 1.0 to 1.6 without loosing essential data and schema-validate the result this should show how in JSON/XML 1.4 and before: this should result in a set of 2 Tool - Component/Service were converted to Tool - to be schema compatible in JSON/XML 1.5 and later: Component/Service were NOT converted to Tool I think this is done, as I see the auto-generated tests and snapshots from get_bom_with_tools_with_component_and_service Bump. this idea of mutual exclusive properties is true for the schema, but not for the data models. therefore, it was removed via 59b0987af61718a1c412aca1b83ca7b3b7b67bae I am trying to go through the changes made, but I'm not understanding where the code is. I see a bunch of code was moved to cyclonedx/serialization/__init__.py in this commit: https://github.com/CycloneDX/cyclonedx-python-lib/commit/376dfa8c1fe59f983ec1217ad83b3cf2bd2e9e1b but when I pull down my branch: https://github.com/jkugler/cyclonedx-python-lib/blob/561_add_components_and_services/cyclonedx/serialization/init.py I don't see that code. Was that code not pushed to my branch? Sorry for the confusionl. I am trying to go through the changes made, but I'm not understanding where the code is. I see a bunch of code was moved to cyclonedx/serialization/__init__.py in this commit: 376dfa8 but when I pull down my branch: https://github.com/jkugler/cyclonedx-python-lib/blob/561_add_components_and_services/cyclonedx/serialization/init.py I don't see that code. Was that code not pushed to my branch? Sorry for the confusionl. i am very sorry for the inconvenience. 376dfa8c1fe59f983ec1217ad83b3cf2bd2e9e1b resp 427add4e8b91a56dba5d565f7809c998c1ba3e96 was a version where i tried to move ToolsRepositoryHelper the helper to place where other existing helpers were. this did not work properly, as it caused cyclic includes. So it was reverted via 4a2ac6526d0cdcb1d01f32d698d8b568399b6454 maybe this helps to see the actual changes: git diff 73007f84fc043924f65560e143ba5adbdab56be2...c937f215e3ac4b56c33cc5da2b0444ee0a22807c https://github.com/jkugler/cyclonedx-python-lib/compare/73007f84fc043924f65560e143ba5adbdab56be2...c937f215e3ac4b56c33cc5da2b0444ee0a22807c I noticed the test coverage for cyclonedx/model/tool.py went from 100% to 97% after the recent changes. Would you like me to make sure the missed statements are checked? I noticed the test coverage for cyclonedx/model/tool.py went from 100% to 97% after the recent changes. Would you like me to make sure the missed statements are checked? did that. please review did that, cov at 100% now. please review Looks great! Let's get it merged! Thanks again for all your help and patience on this.
gharchive/pull-request
2024-06-18T00:01:48
2025-04-01T06:36:52.178616
{ "authors": [ "jkowalleck", "jkugler" ], "repo": "CycloneDX/cyclonedx-python-lib", "url": "https://github.com/CycloneDX/cyclonedx-python-lib/pull/635", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
382048997
这个可以配合databinding 一起来使用吗 有那种bean数据变化 Describe the bug A clear and concise description of what the bug is. To Reproduce Steps to reproduce the behavior: Go to '...' Click on '....' Scroll down to '....' See error Expected behavior A clear and concise description of what you expected to happen. Screenshots If applicable, add screenshots to help explain your problem. Additional context Add any other context about the problem here. 尝试在历史问题搜索答案。 尝试阅读文档找到答案。 尝试阅读Demo找到答案。 尝试自己检查或试验以找到答案。 尝试阅读源代码以找到答案。 请勿将产品的一些特殊交互需求 和 该库暂不支持作为bug混为一谈,请您仔细甄别 如果以上都尝试过了请提一个新的issues 参考提問的智慧 如果还是没有找到答案,提问请带上这几个必要信息 当前使用的版本号 复现操作描述 使用代码 crash日志 gif复现效果 抽取demo 将你出现的问题代码抽出来成一个可直接运行的项目。(最好fork本库修改) 在本地修改demo,然后把commit push到github上,在issue里贴下demo的地址。 有详细的描述才能使得我们更快速的定位问题并解决问题,感谢配合! 有demo可以参考
gharchive/issue
2018-11-19T04:03:23
2025-04-01T06:36:52.213136
{ "authors": [ "CymChad", "to686312" ], "repo": "CymChad/BaseRecyclerViewAdapterHelper", "url": "https://github.com/CymChad/BaseRecyclerViewAdapterHelper/issues/2531", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1110232311
作者已经不维护了,一般使用没什么大问题,有问题的可以试试BRV,作者一直在维护,有问必答 kotlin想坚持使用的可以参考#3561 BRV比较适合databinding,一直在维护,作者解答很快,可以试试😉 使用文档也很全,和作者的net框架一起使用是真的方便快捷😁 适合懒人,但是刷新 内部依赖了 SmartRefreshLayout 就感觉臃肿了些 适合懒人,但是刷新 内部依赖了 SmartRefreshLayout 就感觉臃肿了些 还好,我感觉智能刷新挺好的,要不我也是单独用那个,而且这个扩展了智能刷新布局,刷新加载回调也方便了,还有自动缺省页和分页,我比较看好他集成的功能😁 @dingyong666 还有我目前既在用BRV又在用BRVAH,这个当作adapter,那个就当作RecyclerView来用,我喜欢各取所需😁 @SaltedFish-Extreme 可以封装个接口出来,可以其他 下拉刷新 加载更多 实现接口 灵活替换,因为可能个人喜好用 swiperefresh 等等 其他的,一个项目 多个 通作用的 会变得 多余 @SaltedFish-Extreme 可以封装个接口出来,可以其他 下拉刷新 加载更多 实现接口 灵活替换,因为可能个人喜好用 swiperefresh 等等 其他的,一个项目 多个 通作用的 会变得 多余 封装太麻烦了😂以我目前的水平能把别人的框架用明白就不错了,我只想实现喜欢的效果,就是复制粘贴的UI😂我只想用最少的代码实现最好的效果,功能相同的库多余是没错,但是他的代码写法不是我想要的,有一部分跟我的观念不合,就要用觉得合适的替换掉,我是这么想的,话说老哥都这么晚了,早点睡吧😂 你标题写错了, 我写的叫BRV, 谢谢 🤥 适合懒人,但是刷新 内部依赖了 SmartRefreshLayout 就感觉臃肿了些 SmartRefreshLayout本身就是目前最灵活的下拉刷新框架. 也支持自定义实现. 而且也是属于开箱即用. 我不认为这属于臃肿 另外SmartRefreshLayout本身就支持SwipeRefresh的效果 你标题写错了, 我写的叫BRV, 谢谢 🤥 啊,不好意思,推广都把名字写错了😂 还是谢谢你的推广的! 这个BRVAH的确是烂尾了. 400多个问题很夸张了 在这里推广自己的框架...Hmm 有意思。 @Neo-Turak 这不是看不维护了方便大家嘛
gharchive/issue
2022-01-21T09:15:55
2025-04-01T06:36:52.219176
{ "authors": [ "Neo-Turak", "SaltedFish-Extreme", "dingyong666", "liangjingkanji" ], "repo": "CymChad/BaseRecyclerViewAdapterHelper", "url": "https://github.com/CymChad/BaseRecyclerViewAdapterHelper/issues/3595", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2417811058
node模块怎么删了呀 @limuyang2 大佬,node模块怎么不见了,感觉对一些复杂的设计还是有这种需求的 我复议,在v4实现node较为困难,没法快速开发出来,我只能先用v3了ಥ_ಥ 我看了一下文档,确实没有node了
gharchive/issue
2024-07-19T03:43:57
2025-04-01T06:36:52.220794
{ "authors": [ "dixtdf", "lylelqc", "zhaoxiuyu" ], "repo": "CymChad/BaseRecyclerViewAdapterHelper", "url": "https://github.com/CymChad/BaseRecyclerViewAdapterHelper/issues/3879", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1785181194
🛑 Mi indicador is down In 5a0d880, Mi indicador (https://mindicador.cl) was down: HTTP code: 0 Response time: 0 ms Resolved: Mi indicador is back up in 490919a.
gharchive/issue
2023-07-03T02:49:03
2025-04-01T06:36:52.223567
{ "authors": [ "fbarriosCL" ], "repo": "CymaSuite/upptime", "url": "https://github.com/CymaSuite/upptime/issues/1186", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
33469469
Optional monitors This PR introduces optional monitors feature. The whole idea is that Object doesn't contain __monitor field anymore. TypeInfo_Class will hold a monitor offset, if a class has one (by marking class declaration with @monitor attribute), or 0. Monitor lookup is done with a hash map, if monitorOffset is 0. The hash map is protected by a primitive RW spin lock. Monitor finalization will only lookup for monitors, if a monitor was allocated at least once for the type of object being finalized. druntime counterpart: https://github.com/D-Programming-Language/druntime/pull/789 discussion: http://forum.dlang.org/thread/xpliectmvwrwthamquke@forum.dlang.org I'm in favor of this. The current backward-compatible approach is the way to go, though I agree with @MartinNowak in the long run we may envision complete deprecation. Please rebase and let's push this through. Thanks! Half a year later w/o merging this patch the idea still looks sexy to me. @MartinNowak any plans to move on this in any of 2 competing plans? Well what's going on? "All checks have failed" Half a year later w/o merging this patch the idea still looks sexy to me. @MartinNowak any plans to move on this in any of 2 competing plans? Yes, I'm still opposed to adding a global hash and the outlined plan still makes sense. Furthermore having monitor support on all classes creates ownership/attribute issue for the monitor (https://github.com/MartinNowak/phobos/commit/8cf0ec29ad65ac2a13bd6917b4ff3da0fdea5ab0#diff-4e008aedb3026d4a84f58323e53bf017R4883). I don't have the capacity to pull this story, but starting by adding an @(Object.Monitor) UDA and recognizing that in the compiler should be fairly trivial (e.g. look at the objective-c changes). recognize monitor UDA and deprecate synchronizing on classes without it Let's proceed with stage 1 for when 2.071 opens then? Do we have available documentation / rationale on the deprecated features page? Or a DIP? Do we have available documentation / rationale on the deprecated features page? Or a DIP? Sure, a small entry on http://dlang.org/deprecate.html would be nice, a DIP is overkill though.
gharchive/pull-request
2014-05-14T07:54:21
2025-04-01T06:36:52.233659
{ "authors": [ "DmitryOlshansky", "MartinNowak", "andralex", "ibuclaw", "yglukhov" ], "repo": "D-Programming-Language/dmd", "url": "https://github.com/D-Programming-Language/dmd/pull/3547", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
51102090
[REG2.067a] Issue 13775 - Broken explicit casting of dynamic array slices of known size to static array of different type https://issues.dlang.org/show_bug.cgi?id=13775 Support reinterpret-cast from a CT-known boundaries slice expression that can be implicitly typed as T[n] to another static array type U[m], iff their sizes are same (T[n].sizeof == U[m].sizeof). Another regression filed against this change: https://issues.dlang.org/show_bug.cgi?id=14582
gharchive/pull-request
2014-12-05T14:06:01
2025-04-01T06:36:52.235786
{ "authors": [ "9rnsr", "CyberShadow" ], "repo": "D-Programming-Language/dmd", "url": "https://github.com/D-Programming-Language/dmd/pull/4193", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
56610002
[WIP] std.file Refactoring DO NOT MERGE So far this cleans up DirEntry by factoring out the common interface between Windows and Posix. It also moves the examples in the documentation into unittest blocks instead. There is still more work to be done in std.file, but this is just the start, and I mostly just want to make sure I didn't break anything while doing it. Any news? Going to close as it seems stuck.
gharchive/pull-request
2015-02-05T00:30:41
2025-04-01T06:36:52.237369
{ "authors": [ "DmitryOlshansky", "Orvid" ], "repo": "D-Programming-Language/phobos", "url": "https://github.com/D-Programming-Language/phobos/pull/2962", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
2080206085
生成阶段错误 提示TypeError hexo 7.0.0 提示错误如图 我没有开发过hexo主题 我不知道怎么处理问题 顺便一提 关于live2d的那个js我之后整个包吧 之后用unpkg jsdelivr经常抽风 fix in a88d3ea344177f2621f8240b1b3127c38736f653
gharchive/issue
2024-01-13T09:39:25
2025-04-01T06:36:52.238749
{ "authors": [ "D-Sketon", "MliKiowa" ], "repo": "D-Sketon/hexo-theme-reimu", "url": "https://github.com/D-Sketon/hexo-theme-reimu/issues/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
275542478
File responses should have a filehash object instead of a string. The current API (v1) hard-codes MD5 as the file hash. MD5 should be considered broken as collision attacks are relatively easy to do and do not take long. To make the API easier to update/extend in the future (without even bumping the API version) having filehash be an object would be a lot nicer. filehash: { md5: "abcd", sha256: "1234", sha3-512: "5689" } By just providing multiple hashes adding better hashes is easily doable, without breaking users of the API, and allowing nice updates. In case some hash is severely broken in the future one could even just drop that one from a new version of the API. A good suggestion - will discuss within the team. Fantastic idea, allows future updates easily - will get this implemented. Also while MD5 is unreliable due to collisions, this feature is primarily there as an integrity check, but we will look to adding stronger methods. Yes, all of MD5 and the SHA family of hashes are just to verify integrity of the file transfer (one could also use CRC for that if one wanted to). About something stronger that is the main point of #31.
gharchive/issue
2017-11-21T00:13:03
2025-04-01T06:36:52.263194
{ "authors": [ "Patskimoto", "intenscia", "leper" ], "repo": "DBolical/modioSDK", "url": "https://github.com/DBolical/modioSDK/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
231359871
Add medicaid coverage to clinic data? Thanks for creating an issue! Please fill out this form so we can be sure to have all the information we need, and to minimize back and forth. What are we trying to do? Baltimore is interested in tracking medicaid info for clinics. We should figure out how to implement and what they're hoping to get out of it. What feature or behavior is this required for? ??? How could we solve this issue? (Not knowing is okay!) Probs start by talking to the baltimore folks about how they use this up the road! Anything else? not yet! @lwaldsc or @nerdygirl537 - can you follow up on the user requirements for Baltimore and medicaid, please? This issue doesn't have enough info to do any UX work yet Re: Annie at BAF "So I think what we'd like is the ability to, in DARIA, put in info re: which clinics take Maryland Medicaid and to what gestational age (similar to how you can fill in costs now). Colin mentioned something about creating a clinic lookup tool, so I just wanted to make sure that if/when that is built, it also includes information for Medicaid patients so our CMs can steer them to the best clinic for them." Pretty simple ask here, I think. Based on the above from Annie I think there are two enhancements here: deal with Medicaid clinics in #978 implement something similar to the 'Show only clinics that accept NAF funds' checkbox for Medicaid clinics Either way, the route to both of these is slapping an extra boolean field for accepts_medicaid on the clinic object, and then adjusting the form and tests. Retagging as backend for now, and we can take advantage of the benefits down the road. I'll work on this tonight!
gharchive/issue
2017-05-25T14:42:46
2025-04-01T06:36:52.273778
{ "authors": [ "colinxfleming", "lwaldsc", "mebates", "rudietuesdays" ], "repo": "DCAFEngineering/dcaf_case_management", "url": "https://github.com/DCAFEngineering/dcaf_case_management/issues/1077", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
344863453
task4:This video is unavailable.(deleted by user) when using code to download data from Youtube, I found that part of dataset had been deleted by users. How can I find them? Thanks for your interest in task 4. As explained in the dataset readme: "The script produces missing_files[dataset].csv log files (were [dataset] corresponds to the name of a particular set) with a list of audio files that were not downloaded by the script. After completion if some of the files where not downloaded you might want to run the script a second to download missing files. If you are experiencing problems downloading the full dataset please contact the task organizers (see also task 4 official page)" If you send these files to us (Nicolas Turpault or Romain Serizel), we'll take care of providing the missing audio files. Hello , I am experiencing problems downloading the full dataset (some eval files are missing) , can you provide them ?
gharchive/issue
2018-07-26T14:16:30
2025-04-01T06:36:52.276432
{ "authors": [ "16in17", "rserizel" ], "repo": "DCASE-REPO/dcase2018_baseline", "url": "https://github.com/DCASE-REPO/dcase2018_baseline/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
560549999
DAI searching is completely broken! Nothing happens when you do a normal DAI search... thanks. Note that this doesn't work any better when logged in. This has been resolved.
gharchive/issue
2020-02-05T18:26:12
2025-04-01T06:36:52.287639
{ "authors": [ "cneidle", "kwasiopoku" ], "repo": "DCS-LCSR/ASL-DAI", "url": "https://github.com/DCS-LCSR/ASL-DAI/issues/203", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
718624280
Set default occurrence for all signs being added to signbank if occurrence is the only sign variant Set default occurrence for all signs being added to signbank if occurrence is the only sign variant Done
gharchive/issue
2020-10-10T13:51:08
2025-04-01T06:36:52.288503
{ "authors": [ "kwasiopoku" ], "repo": "DCS-LCSR/ASL-DAI", "url": "https://github.com/DCS-LCSR/ASL-DAI/issues/251", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1050973584
my advanced issue one I am an advanced issue closed
gharchive/issue
2021-11-11T13:20:27
2025-04-01T06:36:52.300011
{ "authors": [ "DEJENERUTA" ], "repo": "DEJENERUTA/my-acme-repository", "url": "https://github.com/DEJENERUTA/my-acme-repository/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
564648799
Add email for all choices being rejected Context When a user receives a rejection and all of their course choices have been a rejected they should be sent an email. Changes proposed in this pull request Add an email to the CandidateMailer Guidance to review Emails https://docs.google.com/document/d/1VH_nxuLTiVCkmKhC4eaBpFC61Hd5_Ol1tKwOul-2uUs/edit#heading=h.hz139gcneav5 This email needs a provider name, course name, candidate name, amount of choices (for pluralisation and rejection reason. I've struggled to get all the below objects associated using build_stubbed. Is there an easier way? The current implementation is a pain for adding it to PreviewCandidateMailer Also, what would happen if a provider refused to give feedback was not covered. I added the following: This needs a content review. Link to Trello card https://trello.com/c/22e7C80p/840-email-🙅♀️-a-provider-has-rejected-your-application-to-candidate Things to check [x] This code doesn't rely on migrations in the same Pull Request [x] If this code includes a migration adding or changing columns, it also backfills existing records for consistency [x] API release notes have been updated if necessary [x] New environment variables have been added to the Azure config I had a similar struggle with build_stubbed yesterday - it has limitations! I was able to set up a one-many association by using a combination of hard-coded ids and adding to collections using plain build (so a bit hacky). https://github.com/DFE-Digital/apply-for-postgraduate-teacher-training/blob/master/spec/mailers/candidate_mailer_spec.rb#L287-L298 It may be completely different for your case.
gharchive/pull-request
2020-02-13T12:25:06
2025-04-01T06:36:52.315698
{ "authors": [ "davidgisbey", "stevehook" ], "repo": "DFE-Digital/apply-for-postgraduate-teacher-training", "url": "https://github.com/DFE-Digital/apply-for-postgraduate-teacher-training/pull/1358", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
536242648
Minor API documentation fixes Context This is a follow-up to https://github.com/DFE-Digital/apply-for-postgraduate-teacher-training/pull/826. A number of issues were raised after manual testing by @fofr . This PR addresses the remaining fixes to API docs. Changes proposed in this pull request [x] Work experience commitment example and description needed (already done) [x] Qualification examples are missing the "institution_name" and "equivalency" examples (already done) [x] We should include an example of an ISO8601 date in the since param (PH actually caused an error with this when he tried it) - (done in earlier PR) [x] reference content should not be marked "optional" bc we only send apps with references - changed the wording to emphasise that content is required to send apps to providers [x] Clarify that tokens can only be passed via header, not via URL param - added a bit of extra copy to clarify I've not tried to add the state diagram (suggestion from @duncanjbrown) in this PR because there isn't an obvious place to put it in the docs (without causing a confusion). This one is open to debate! Guidance to review Do the copy changes make sense? Link to Trello card 1176 - Minor API fixes following user testing with Paul H Env vars [x] No env vars Perhaps the state diagram could go in the general introduction? On the basis that it's a useful high-level summary of how the thing works. We might need to prune it a bit. @duncanjbrown I've added a section to the api-docs home page called 'Application Lifecycle' with a copy of the state diagram. Wording is all up for debate. I've not tried to trim the diagram (it's just automatically copied from the generate_state_diagram rake task into the public directory). Perhaps the state diagram could go in the general introduction? On the basis that it's a useful high-level summary of how the thing works. We might need to prune it a bit. @duncanjbrown I've added a section to the api-docs home page called 'Application Lifecycle' with a copy of the state diagram. Wording is all up for debate. I've not tried to trim the diagram (it's just automatically copied from the generate_state_diagram rake task into the public directory). :+1: I think this is useful
gharchive/pull-request
2019-12-11T09:24:25
2025-04-01T06:36:52.322342
{ "authors": [ "duncanjbrown", "stevehook" ], "repo": "DFE-Digital/apply-for-postgraduate-teacher-training", "url": "https://github.com/DFE-Digital/apply-for-postgraduate-teacher-training/pull/849", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1947254506
ServiceDelivery None Clustered Index Added ServiceDelivery None Clustered Index Added out of date
gharchive/pull-request
2023-10-17T12:04:14
2025-04-01T06:36:52.328511
{ "authors": [ "MartinBelton-gov", "lazcool" ], "repo": "DFE-Digital/fh-service-directory-api", "url": "https://github.com/DFE-Digital/fh-service-directory-api/pull/143", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
415539963
Add course scope for opted in providers Context The courses endpoint should only return courses where their providers are opted in Changes proposed in this pull request Add scope to courses endpoint Guidance to review /api/v1/courses Ensure courses are scoped by providers that have opted in
gharchive/pull-request
2019-02-28T10:02:36
2025-04-01T06:36:52.336795
{ "authors": [ "dankmitchell" ], "repo": "DFE-Digital/manage-courses-backend", "url": "https://github.com/DFE-Digital/manage-courses-backend/pull/156", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2052188889
[6527] Enable performance/profile monitoring Context Enables Sentry's performance and profile monitoring so we can setup some anomaly alerts around this category. Start with a low traces_sample_rate rate. We'll assess how noisy it is in Sentry and adjust accordingly. Set up anomaly alerts in register support slack channel
gharchive/pull-request
2023-12-21T10:50:19
2025-04-01T06:36:52.341474
{ "authors": [ "darokel" ], "repo": "DFE-Digital/register-trainee-teachers", "url": "https://github.com/DFE-Digital/register-trainee-teachers/pull/3883", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
891097947
remove-paas-url-from-tfvar-file Use new PaaS credentials Jira ticket URL Just add the ticket number to the end: https://dfedigital.atlassian.net/browse/TEVA- Changes in this PR: Is there anything specific you want feedback on? Screenshots of UI changes: Before After Next steps: [ ] Terraform deployment required? [ ] New development configuration to be shared? Review app deployed to https://teaching-vacancies-review-pr-3462.london.cloudapps.digital
gharchive/pull-request
2021-05-13T14:37:49
2025-04-01T06:36:52.347764
{ "authors": [ "felixtheflex", "twd-tv-ci" ], "repo": "DFE-Digital/teaching-vacancies", "url": "https://github.com/DFE-Digital/teaching-vacancies/pull/3462", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
899503907
Do not refer to ActionMailer::Base in initializers This causes issues with autoloading (invoking behaviour that has been deprecated in Zeitwerk when certain stars align, in our case if we add ActionText, c.f. https://github.com/rails/rails/issues/36546) Instead of adding configuration to ActionMailer::Base, set it on Rails.configuration.action_mailer instead. Review app deployed to https://teaching-vacancies-review-pr-3521.london.cloudapps.digital
gharchive/pull-request
2021-05-24T09:53:52
2025-04-01T06:36:52.350441
{ "authors": [ "csutter", "twd-tv-ci" ], "repo": "DFE-Digital/teaching-vacancies", "url": "https://github.com/DFE-Digital/teaching-vacancies/pull/3521", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
271054219
Refine report 2 Ensure each plot is on a new page Consider making plots landscape Decided not to progress
gharchive/issue
2017-11-03T17:18:00
2025-04-01T06:36:52.354796
{ "authors": [ "adamrobinson361", "isi-avbulimen" ], "repo": "DFEAGILEDEVOPS/schools-workforce-benchmarking", "url": "https://github.com/DFEAGILEDEVOPS/schools-workforce-benchmarking/issues/33", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
210820070
Not compatible with Semanticolor I'm trying to use code-peek 1.4.16 with Semanticolor, but there's a conflict. The problem is that to do its highlighting, Semanticolor prefixes the name of the current grammar with "semanticolor", so code-peek gives the error "Peek function does not currently support semanticolor - Ruby files". So...perhaps it should ignore the "semanticolor" prefix or find some other way of determining the language? 1.4.17 will fix this
gharchive/issue
2017-02-28T15:30:54
2025-04-01T06:36:52.357154
{ "authors": [ "DFreds", "marnen" ], "repo": "DFreds/code-peek-atom", "url": "https://github.com/DFreds/code-peek-atom/issues/21", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2740913176
Windows 版本 19042.5198 Snap Hutao 版本 1.12.8.0 设备 ID BE9C5A62B14493ED583B58443DC34433 问题分类 安装和环境 发生了什么? 每次版本升级之后,不管是哪种方式更新,更新完之后第一次打开,如果开着代理,他会弹出“解除loopback”弹窗,然后无论点击解除与否,胡桃许多界面直接变成空白,无法打开,首页也只有启动游戏能够点击,其他均是加载中的界面,而且重启电脑和胡桃都无法解决,只有重装然后第一次运行的时候不开着代理,所有组件才能正常加载。如果需要日志和图片我下次更新的时候补上 你期望发生的行为? 开着代理不影响组件的加载 最后一步 [x] 我认为上述的描述已经足以详细,以允许开发人员能复现该问题 首页也只有启动游戏能够点击,其他均是加载中的界面 在代理中排除 api.snapgenshin.com
gharchive/issue
2024-12-15T21:38:25
2025-04-01T06:36:52.360100
{ "authors": [ "Lightczx", "kissgo" ], "repo": "DGP-Studio/Snap.Hutao", "url": "https://github.com/DGP-Studio/Snap.Hutao/issues/2298", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1657300990
Extract model data at location(s) Narrow API to model extraction. More explicit what attributes that are needed to extract data from a model. I.e. the observed values, color, filename etc. is not relevant in this context. I see the point (no pun intended) from an architectural point of view - but as a user I think it becomes more difficult to understand (more objects/classes). Wouldn't it be better to let Point and Track be a Mixin (or a protocol) and then let the extract method accept anything which is a Point/Track? I see the point (no pun intended) from an architectural point of view - but as a user I think it becomes more difficult to understand (more objects/classes). Wouldn't it be better to let Point and Track be a Mixin (or a protocol) and then let the extract method accept anything which is a Point/Track? I think it is difficult to enforce compliance with an ABC or Protocol, since they only enforce the names of the methods (and names of arguments), but not attributes. So this last change tries to balance simplicity for the user, no need for conversion and still communicating the needs of the extract methods.
gharchive/pull-request
2023-04-06T12:06:43
2025-04-01T06:36:52.378574
{ "authors": [ "ecomodeller", "jsmariegaard" ], "repo": "DHI/fmskill", "url": "https://github.com/DHI/fmskill/pull/181", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2512011122
Text boxes within groupboxes not stretching properly I have the following XAML <w3c:GroupBox Grid.Row="0" Grid.Column="0" Heading="test" HorizontalAlignment="Stretch" VerticalAlignment="Top" MinWidth="300"> <Grid HorizontalAlignment="Stretch" VerticalAlignment="Stretch"> <Grid.RowDefinitions> <RowDefinition Height="Auto" /> <RowDefinition Height="Auto" /> </Grid.RowDefinitions> <Grid.ColumnDefinitions> <ColumnDefinition Width="Auto" /> <ColumnDefinition Width="*" /> </Grid.ColumnDefinitions> <TextBox x:Name="Test" Header="Test ID:" Grid.Column="1" HorizontalAlignment="Stretch" VerticalAlignment="Top" Grid.Row="0" /> <TextBox x:Name="Test2" Header="Test ID:" Grid.Column="1" HorizontalAlignment="Stretch" VerticalAlignment="Top" Grid.Row="1" Margin="0,10,0,0" /> </Grid> </w3c:GroupBox> and the Stretch property is not working properly, as I just see this The textboxes do, however, resize when you type in them Instead of setting the GroupBox property HorizontalAlignment="Stretch" you should set HorizontalContentAlignment="Stretch" That should fix the problem you are seeing.
gharchive/issue
2024-09-07T18:20:11
2025-04-01T06:36:52.383738
{ "authors": [ "DHancock", "cxnky" ], "repo": "DHancock/WinUI3Controls", "url": "https://github.com/DHancock/WinUI3Controls/issues/49", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
1377118774
Create simple component for Relation type icons RelationType dictionary keys in a circle. Something like this Needs scaling - relation of circle size to font size
gharchive/issue
2022-09-18T17:00:35
2025-04-01T06:36:52.410253
{ "authors": [ "Ptrhnk", "adammertel" ], "repo": "DISSINET/InkVisitor", "url": "https://github.com/DISSINET/InkVisitor/issues/1234", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
925592619
SAC with gSDE not working on MinitaurBulletEnv-v0 Hello @araffin, I trained the agent on MinitaurBulletEnv-v0 but without success. I tried to use the Min-max normalization method because every observation has its own range of values, but it's not working and without normalization, I get better results. I test it on my platform RL Toolkit, which is based on TF but is almost the same as yours. Why is MinitaurBulletEnv-v0 not working, when the agent here is learning? My tested hyperparameters: Hyperparameter Value n_timesteps 1000000 learning_rate 0.00073 batch_size 256 buffer_size 1000000 ent_coef auto gamma 0.99 learning_starts 10000 update_interval 64 When you'll find better hyperparameters or anything else please share it with me. https://user-images.githubusercontent.com/74611856/122724806-9606cf00-d274-11eb-8212-6c414012e29e.mp4 Thanks a lot. Hello, So after a quick trial, I found those hyperparameters to be working (for unstructured noise): MinitaurBulletEnv-v0: n_timesteps: !!float 1e6 policy: 'MlpPolicy' learning_rate: !!float 3e-4 buffer_size: 100000 batch_size: 256 ent_coef: 'auto' train_freq: 1 gradient_steps: 1 learning_starts: 10000 with gSDE (the noise sampling frequency is quite small and therefore quite close to unstructured noise): MinitaurBulletEnv-v0: n_timesteps: !!float 1e6 policy: 'MlpPolicy' learning_rate: !!float 3e-4 buffer_size: 1000000 batch_size: 256 ent_coef: 'auto' gamma: 0.99 train_freq: 4 gradient_steps: 4 learning_starts: 10000 use_sde: True policy_kwargs: "dict(log_std_init=-3)" As mentioned in the gSDE paper (the arxiv version will be updated tomorrow: https://arxiv.org/abs/2005.05719 ), the main strength of gSDE is not in simulation but on a real robot (as it reduces wear-and-tear while keeping good performance). I will try to upload the learning curves and trained agent soon. The training reward with unstructured noise: I thought that the MinitaurBulletEnv-v0 is Sim-to-Real problematics and gSDE is useful for that too. I will test it with a higher update freq than 64 and a lower learning rate. Thanks. I thought that the MinitaurBulletEnv-v0 is Sim-to-Real problematics and gSDE is useful for that too. gSDE was designed to run RL directly on real robots (no sim2real) even though it should help for sim2real pb too (however, I'm not sure if MinitaurBulletEnv-v0 is Sim-to-Real problematics or not). I read about Minitaur here. Can you share with me please, Actor loss, Critic loss, Steps (at episode) charts for detailed analysis of my problematics. I try your hyperparameters without significant improvement. I use client-server architecture for training RL agents (with Reverb). Thanks. Best for you would be to do a run using SB3 + the rl zoo: the hyperparams: MinitaurBulletEnv-v0: n_timesteps: !!float 1e6 policy: 'MlpPolicy' learning_rate: !!float 3e-4 buffer_size: 1000000 batch_size: 256 ent_coef: 'auto' gamma: 0.99 train_freq: 4 gradient_steps: 4 learning_starts: 10000 use_sde: True policy_kwargs: "dict(log_std_init=-3)" Training with tensorboard (logging to /tmp/tensorboard_sb3/ here) python train.py --algo sac --env MinitaurBulletEnv-v0 -tb /tmp/tensorboard_sb3/ --num-threads 2 --eval-episodes 20 --n-eval-envs 5 The learning curves (blue is unstructured noise, orange with gSDE): Note that the true performance of SAC is higher because I display the training reward here, not the evaluation one using the deterministic controller. PS: I think that it is because the learner is faster than the agent. When I try for example AntBulletEnv-v0 everything is good, but the agent plays faster than the learner. In this case, the learner has more data to training from. you mean that you are learning in parallel of data collection? Yes this can be definitely a problem. Btw, the new version of the gSDE paper is online https://arxiv.org/abs/2005.05719 ;) A summary of where it is useful: Yes, the learner instance is learning in parallel to data collection. Thanks a lot for your time, I will be trying it with your hyperparameters and try to make the learner slower than the agent is collecting data because those configuration is working on other environments. Yeah, thanks for the answers I think this issue is solved. I must still be working on RL-Toolkit. Expectation over time means the average over the time in continuity cost equation? yes. I'm using a wrapper to compute it: class ContinuityCostWrapper(gym.Wrapper): """ Add continuity cost to the reward. It assumes that the action space is normalized and symmetric (actions in [-1, 1]). :param env: :param weight_continuity: :param verbose: :param print_freq: Print every n episodes the mean continuity cost """ def __init__(self, env: gym.Env, weight_continuity: float = 0.0, verbose: int = 0, print_freq: int = 1): super(ContinuityCostWrapper, self).__init__(env) self.last_action = None self.weight_continuity = weight_continuity self.verbose = verbose self.continuity_hist = [] self.unnormalized_hist = [] self.n_episodes = 0 self.print_freq = print_freq def reset(self): self.last_action = None self.n_episodes += 1 return self.env.reset() def step(self, action): obs, reward, done, info = self.env.step(action) # Continuity cost if self.last_action is not None: max_delta = 2.0 # for the action space: high - low = 1 - (-1) = 2 continuity_cost = np.mean((action - self.last_action) ** 2 / max_delta ** 2) unnormalized_cost = np.mean((action - self.last_action) ** 2) self.continuity_hist.append(continuity_cost) self.unnormalized_hist.append(unnormalized_cost) continuity_cost = self.weight_continuity * continuity_cost self.last_action = action.copy() else: continuity_cost = 0.0 self.last_action = action.copy() if done: continuity_score = 100 * np.mean(self.continuity_hist) if self.verbose > 0 and self.n_episodes % self.print_freq == 0: print(f"n_step={len(self.continuity_hist)}") print(f"Continuity={continuity_score:.5f} +/ {np.std(self.continuity_hist):.5f}") # print(f"Unnormalized continuity={np.mean(self.unnormalized_hist):.5f}\n") info["continuity_score"] = continuity_score reward -= continuity_cost info["continuity_cost"] = continuity_cost return obs, reward, done, info @araffin Thanks all is working now. I must use a TQC algorithm instead of classic SAC, because the learner was faster than the agent collecting experiences. TotalInteractions: 1532116 (the agent steps in the environment) Train step: 999906 (the learner's training steps) Now is the agent about 500 000 steps before the learner and learning process is more accurate. https://user-images.githubusercontent.com/74611856/124826185-d81e5900-df74-11eb-80a2-3b6fb800bdec.mp4
gharchive/issue
2021-06-20T12:48:51
2025-04-01T06:36:52.461214
{ "authors": [ "araffin", "markub3327" ], "repo": "DLR-RM/rl-baselines3-zoo", "url": "https://github.com/DLR-RM/rl-baselines3-zoo/issues/117", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
564632104
Create typing stub files Mypy doc on stub files. https://mypy.readthedocs.io/en/stable/stubs.html Create stub files for used third party libraries that do not provides their own types. Create stub files for gitlab2prov code to allow projects that import modules from gitlab2prov to also import their type signatures. Sub-Issue of #26 Postponed
gharchive/issue
2020-02-13T11:52:00
2025-04-01T06:36:52.463465
{ "authors": [ "cdboer", "onyame" ], "repo": "DLR-SC/gitlab2prov", "url": "https://github.com/DLR-SC/gitlab2prov/issues/28", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
148921701
DMIunipgAI-VacuumCleaner2016-trivelle versione "Trivelle" del Vacuum Cleaner, file della competition 2016 https://github.com/DMIunipg/AI-Project-VacuumEnvironment/pull/2.patch Step 1: From your project repository, check out a new branch and test the changes. git checkout -b valefranz-master master git pull git://github.com/valefranz/AI-Project-VacuumEnvironment.git master Step 2: Merge the changes and update on GitHub. git checkout master git merge --no-ff valefranz-master git push origin master
gharchive/pull-request
2016-04-17T05:55:03
2025-04-01T06:36:52.478306
{ "authors": [ "valefranz" ], "repo": "DMIunipg/AI-Project-VacuumEnvironment", "url": "https://github.com/DMIunipg/AI-Project-VacuumEnvironment/pull/2", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1033696615
File naming It seems that many file names between the requester/responder are the same. Generally we avoid h/c file collisions because they are in different directories. I’m not so sure this applies to object files though as they may be built into the same directory and have the same name if some client wants to use both the requester and responder. When I ported the libspdm to a Windows static library to be finally built as a Windows dll, I ran into problems and had to rename some of the files so their object files would not collide. The VC++ standard project/solution build system normally dumps all object files to a single output dir. I did not see an easy way to change this, so I resolved it by changing the files with the name collision to have “requester” in front of them. I think other potential clients may fun into this problem, and was wondering if we should name the files differently, by prepending requester or responder to then. Example. Instead of “communication.c”, name it “requester_communication.c” and “responder_communication.c”. Resolve together with https://github.com/DMTF/libspdm/issues/155 Resolved by https://github.com/DMTF/libspdm/pull/278
gharchive/issue
2021-10-22T15:19:13
2025-04-01T06:36:52.492547
{ "authors": [ "jyao1", "rkongintel", "steven-bellock" ], "repo": "DMTF/libspdm", "url": "https://github.com/DMTF/libspdm/issues/221", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
268193426
Can't get access to Phone No. Field in Event Enrollment I can't seem to find a token to access the Phone Number field that is built into the Enroll section of an Event. On this page: https://github.com/DNNCommunity/DNN.Events/wiki/Tokens-to-be-used-in-Templates I found the token for the name: [event:signupusername] and Email address [event:signupuseremail] But I have no idea how to access the data in the Phone No. field. -Tom Currently, the Phone number is not avaialble as token. I will add this as an enhancement request Wonderful. Thanks for responding and letting me know. Tom
gharchive/issue
2017-10-24T21:32:55
2025-04-01T06:36:52.501162
{ "authors": [ "EPTamminga", "thosalbert" ], "repo": "DNNCommunity/DNN.Events", "url": "https://github.com/DNNCommunity/DNN.Events/issues/57", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2624951731
Added my notebook with analysis of German credit risk dataset. This includes correlations, a few graphs, and reports of my findings Good job @samc5 10/10
gharchive/pull-request
2024-10-30T18:20:37
2025-04-01T06:36:52.570798
{ "authors": [ "aakankshaduggal", "samc5" ], "repo": "DS219/spark-seprep", "url": "https://github.com/DS219/spark-seprep/pull/115", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2648738456
Is there a Docker deployment solution or a FastAPI server setup available for docling? Question I am looking to deploy docling and would like to know if there is an existing Docker deployment solution or a FastAPI server setup available. ... Hi @ShedrachJonah11, we have a webserver for docling in the works. It is currently experimental stage. See here. There is also a Dockerfile in this repo to demonstrate how to run docling on the container.
gharchive/issue
2024-11-11T09:35:27
2025-04-01T06:36:52.585623
{ "authors": [ "ShedrachJonah11", "cau-git" ], "repo": "DS4SD/docling", "url": "https://github.com/DS4SD/docling/issues/293", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
374200568
Submit zip jobs Adding support for job submission as zip files to twister2. I think we should determine the content of the zip file. What files and directories should there be? We need to unzip it on workers and add jar files to the classpath. As I understand, we expect the user to specify the job type on command line as "job_type=zip" or "job_type=jar". We are also expecting the user to specify the job file on the command line as "job_file=xyz.jar" or "job_file=xyz.zip". I think we may not need the extra job_type parameter. We can just check the extension of the job_file. we should properly pack the unzipped files into the job package. Currently, TarGzipPacker in ResourceAllocator packs all files into a tar.gz file. That file is transferred to the workers. This packer currently packs job description file[job-name.job], user job jar file (jar file specified on command line), and conf directory. We need to add any newly added files or directories from zip file to the job tar.gz package. So, this all depends on the format of the original zip file. What will be the content of that zip file.
gharchive/pull-request
2018-10-26T02:17:45
2025-04-01T06:36:52.601130
{ "authors": [ "Dananji", "ahmet-uyar" ], "repo": "DSC-SPIDAL/twister2", "url": "https://github.com/DSC-SPIDAL/twister2/pull/69", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2559004743
Update links in docs to new Angular site A while ago, Angular moved all of their documentation to a new site (angular.dev). All the links in our docs are now out of date and should be updated. As far as I can see, there are only some references from the frontend testing docs. Those are outdated anyway, we don't actively use playwright for testing anymore, instead we use Storybook.
gharchive/issue
2024-10-01T11:55:04
2025-04-01T06:36:52.602254
{ "authors": [ "MoritzWeber0", "zusorio" ], "repo": "DSD-DBS/capella-collab-manager", "url": "https://github.com/DSD-DBS/capella-collab-manager/issues/1862", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2421949435
🛑 Fakultas Kedokteran Gigi is down In 861439b, Fakultas Kedokteran Gigi (https://dent.unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: Fakultas Kedokteran Gigi is back up in 2e803fe after 26 minutes.
gharchive/issue
2024-07-22T05:10:42
2025-04-01T06:36:52.607067
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/10180", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2623316939
🛑 Unhas Main is down In 83295c7, Unhas Main (https://unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: Unhas Main is back up in 9a251f3 after 5 minutes.
gharchive/issue
2024-10-30T08:54:11
2025-04-01T06:36:52.609426
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/11728", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2062753831
🛑 SSO Unhas is down In 138edc9, SSO Unhas (https://sso.unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: SSO Unhas is back up in d8f6f3a after 5 minutes.
gharchive/issue
2024-01-02T17:52:10
2025-04-01T06:36:52.611854
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/5538", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2195830961
🛑 Fakultas Kedokteran Gigi is down In 8019097, Fakultas Kedokteran Gigi (https://dent.unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: Fakultas Kedokteran Gigi is back up in 75da26c after 4 minutes.
gharchive/issue
2024-03-19T19:34:54
2025-04-01T06:36:52.614510
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/7208", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2216970700
🛑 Fakultas Vokasi is down In 25d8c10, Fakultas Vokasi (https://vokasi.unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: Fakultas Vokasi is back up in ab4a137 after 10 minutes.
gharchive/issue
2024-03-31T13:18:24
2025-04-01T06:36:52.617200
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/7522", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2221445946
🛑 Unhas Main is down In 7810258, Unhas Main (https://unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: Unhas Main is back up in 88fc53d after 11 minutes.
gharchive/issue
2024-04-02T21:07:06
2025-04-01T06:36:52.619596
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/7622", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2752928581
Shift Register. Please review modified: document/rtl/modules.md new file: document/rtl/shift_reg.md new file: document/rtl/shift_reg_top.svg new file: source/shift_reg.sv Rejected. Unnecessary addition
gharchive/pull-request
2024-12-20T15:11:24
2025-04-01T06:36:52.621970
{ "authors": [ "SubhanBihan", "foez-ahmed" ], "repo": "DSi-DV/maverickOne", "url": "https://github.com/DSi-DV/maverickOne/pull/78", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
345840781
[DS-3971] BitstreamStorageServiceImpl: Do not call update If the BitstreamStorageService calls update on the just cloned bitstream, this bitstream will not belong to any Item and the user may get authorization failures. The update call can be removed here, because the clone method is only used by the AbstractVersionProvider and it will call update on the bitstream later by itself. https://jira.duraspace.org/browse/DS-3971 I dislike having BitstreamStorageServiceImpl make changes to an entity and then give up control of it without persisting them, but we may have no choice here. If so, the unpersisted state should be documented. The place for that is in BitstreamStorageService#clone, which is not documented at all. @mwoodiupui I just added some comment about the missing update() call to the existing comment of BitstreamStorageService#clone. Is this what you had in mind? Yes, thank you. Whenever we make changes to an entity which we do not commit, the caller should be warned. I tested it, a submitter can create a new version of its item! Just a thought: Instead of changing the cloned bitstream in the BitstreamStorageServiceImpl#clone, does anything speak against moving all changes to BitstreamServiceImpl#clone? In this way, there will not be a unpersisted state. We have created a pull request which avoid an unpersisted state in BitstreamStorageServiceImpl#clone https://github.com/DSpace/DSpace/pull/2428#issue-279153056 Closing, replaced by #2428 (Please review/test that PR so that we can get it resolved/merged)
gharchive/pull-request
2018-07-30T16:44:05
2025-04-01T06:36:52.626347
{ "authors": [ "AlexanderS", "marsaoua", "mwoodiupui", "tdonohue" ], "repo": "DSpace/DSpace", "url": "https://github.com/DSpace/DSpace/pull/2139", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
54029636
[DS-2345] Fixes wrong start index starting at 0 instead of 1 for OpenSearch This fixes the bug for the Discovery implementation. I think for the StandardOpenSearchGenerator (old Lucene based) implemenatation there are 2 lines which should be changed too. Line 72 and Line 114. But we are using Discovery which makes it harder to test it. I accidentally merged the changes on dspace master to the Branch [DS-2345]. Not sure how this affects the pull request. Don't worry, we have a way to get just the stuff we need. Here is more Information on the bug: If you look at the result xml file of this query: http://demo.dspace.org/xmlui/open-search/discover?format=kops&query=author%3A*&start=0 it says that start index is 1 <opensearch:startIndex>1</opensearch:startIndex> This query says the same: http://demo.dspace.org/xmlui/open-search/discover?format=kops&query=author%3A*&start=1 but the result is different (starting with the second result of the first query) The expected behaviour would be: the same query result for both querys because OpenSearch starts counting at 1 not at 0 and if <opensearch:startIndex>1</opensearch:startIndex> is equal in both cases the result has to be the same too. +1 tested, cherry-picked
gharchive/pull-request
2015-01-12T08:27:10
2025-04-01T06:36:52.630383
{ "authors": [ "christian-scheible", "helix84" ], "repo": "DSpace/DSpace", "url": "https://github.com/DSpace/DSpace/pull/824", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
306232079
com.android.builder.dexing.DexArchiveBuilderException com.android.builder.dexing.DexArchiveBuilderException on Android Issue Type [x] Bug Report [ ] Feature Request Description This Gradle exception occurs when trying to build app with JDA https://developer.android.com/studio/write/java8-support.html @natanbc thank you
gharchive/issue
2018-03-18T11:24:50
2025-04-01T06:36:52.637410
{ "authors": [ "gocrafterlp", "natanbc" ], "repo": "DV8FromTheWorld/JDA", "url": "https://github.com/DV8FromTheWorld/JDA/issues/649", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
849194708
First pass on stage channels Pull Request Etiquette [x] I have checked the PRs for upcoming features/bug fixes. [x] I have read the contributing guidelines. Changes [x] Internal code [x] Library interface (affecting end-user code) [ ] Documentation [ ] Other: _____ Closes Issue: #1572 Description This adds support for stage channels, which are basically just voice channels with topics. Freezing this for now due to discord/discord-api-docs#2751 Please note that this feature is still under active development (even though it is launched), and thus there is higher than normal risk of things possibly changing. The API for stage channels is kinda weird. I did some digging and found out a few issues that will be annoying to deal with: The request to speak endpoint seems to be REST-only (so you can't join and immediately speak) The endpoint also requires you to specify which channel you are currently in (which requires caching, not a problem for us) They expect an ISO timestamp for the request to speak (i can't understand why but ok) You can suppress yourself? I have no idea why this is a thing Apparently, bots can also approve requests to speak, this is limited through MUTE_MEMBERS permission The request to speak timestamp is also part of voice states, so we can derive an event there Can't see any methods to invite someone to speak, I think that's possible in stage channels (works like speaking requests but backward: the user has to accept the invite from a moderator to speak) Any updates? Can't see any methods to invite someone to speak, I think that's possible in stage channels (works like speaking requests but backward: the user has to accept the invite from a moderator to speak) The bot I use (which uses this JDA lib) can have a moderator role and bypass any invites?? It will be ready when it is ready. Stage channels are still in flux behind the scenes in the Discord API. Should we move the speaker methods such as approveSpeaker and inviteSpeaker into the Member interface instead? Feature TODO [ ] Stage Instance events [ ] Stage Instance moderation (set topic etc) [ ] Voice states for lurkers Overall TODO [ ] Documentation [ ] Checks [ ] Testing
gharchive/pull-request
2021-04-02T12:38:21
2025-04-01T06:36:52.646383
{ "authors": [ "DV8FromTheWorld", "MinnDevelopment", "TheChilliPL", "jzvi12" ], "repo": "DV8FromTheWorld/JDA", "url": "https://github.com/DV8FromTheWorld/JDA/pull/1575", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
62729672
Bug : More paragraphs in comments Hi First, your DZReadability is powerfull ! I've seen juste one bug, its when the comments bloc have more than article. By example for this news : http://blog.lefigaro.fr/football/bruno_roger-petit/2015/03/subasic-heureux-pour-monaco-peut-on-se-rejouir-davoir-ete-ridicule.html // count how many p tags are inside the parent NSArray *pNodes = [parent nodesMatchingSelector:@"p"]; Did you have an idea for fix it please ? :) Thx Realbug Hi Thank you. That's something that I will try to focus on in the next release. I think of adding the ability to provide tag ids and classes for instant win to the algorythm before start. It will help to identify article block on specific sites and may help cure these errors with finding the right article block. I will also implement ignoring comment blocks by their most common id and class names. I don't know when I'll return to this project for the next iteration. Any help is appreciated. @DZamataev cant able to get the heading of the web page , i have tried with various Readability Parser Option but no result , can u help me ? @praveengodz Sure, give me the url you are trying to parse please. @DZamataev URL : http://trak.in/tags/business/2016/06/14/employees-startup-must-have/ From this url except main heading (title) other are so perfect . Thanks for your support Denis.
gharchive/issue
2015-03-18T16:03:17
2025-04-01T06:36:52.652600
{ "authors": [ "DZamataev", "RealBug", "praveengodz" ], "repo": "DZamataev/DZReadability", "url": "https://github.com/DZamataev/DZReadability/issues/2", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }