Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Workaround dist.apache.org download restrictions #752

Merged

Conversation

gerlowskija
Copy link
Contributor

@gerlowskija gerlowskija commented Jan 23, 2025

After a "passing" RC, the wizard has RMs download the CRDs and helm charts from a 'staging' area on dist.apache.org and then upload them to the final location. We have scripts to do this, but these were broken recently when dist.apache.org changed its robots.txt to disallow unknown "crawlers".

This commit gets our scripting working again by tweaking a wget invocation to not strictly obey the robots.txt for dist.apache.org, which likely isn't intended for restricting foundation-internal usecases such as ours.

After a "passing" RC,the wizard has RMs download the CRDs and helm
charts from a 'staging' area on dist.apache.org and then upload them to
the final location.  We have scripts to do this, but these were broken
recently when dist.apache.org changed its robots.txt to disallow unknown
"crawlers".

This commit gets our scripting working again by tweaking a `wget`
invocation to not strictly obey the robots.txt for dist.apache.org,
which likely isn't intended for restricting foundation-internal usecases
such as ours.
@gerlowskija gerlowskija merged commit 0ffddb6 into apache:main Jan 24, 2025
1 check passed
@gerlowskija gerlowskija deleted the ignore-robots-for-single-dir-downloads branch January 24, 2025 19:06
gerlowskija added a commit that referenced this pull request Jan 24, 2025
After a "passing" RC,the wizard has RMs download the CRDs and helm
charts from a 'staging' area on dist.apache.org and then upload them to
the final location.  We have scripts to do this, but these were broken
recently when dist.apache.org changed its robots.txt to disallow unknown
"crawlers".

This commit gets our scripting working again by tweaking a `wget`
invocation to not strictly obey the robots.txt for dist.apache.org,
which likely isn't intended for restricting foundation-internal usecases
such as ours.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants