2020-10-10 14:16:11 +02:00
|
|
|
# typed: false
|
2019-04-19 15:38:03 +09:00
|
|
|
# frozen_string_literal: true
|
|
|
|
|
2016-07-12 19:46:29 +01:00
|
|
|
require "open3"
|
2016-06-03 13:05:18 +01:00
|
|
|
|
2021-03-16 18:15:21 +01:00
|
|
|
require "extend/time"
|
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
module Utils
|
|
|
|
# Helper function for interacting with `curl`.
|
|
|
|
#
|
|
|
|
# @api private
|
|
|
|
module Curl
|
2021-03-16 18:15:21 +01:00
|
|
|
using TimeRemaining
|
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
module_function
|
|
|
|
|
|
|
|
def curl_executable
|
|
|
|
@curl ||= [
|
|
|
|
ENV["HOMEBREW_CURL"],
|
|
|
|
which("curl"),
|
|
|
|
"/usr/bin/curl",
|
|
|
|
].compact.map { |c| Pathname(c) }.find(&:executable?)
|
2021-01-24 21:40:41 -05:00
|
|
|
raise "No executable `curl` was found" unless @curl
|
2020-10-10 15:23:03 +02:00
|
|
|
|
|
|
|
@curl
|
|
|
|
end
|
2016-06-03 13:05:18 +01:00
|
|
|
|
2020-12-19 17:56:25 -05:00
|
|
|
def curl_args(*extra_args, **options)
|
2020-10-10 15:23:03 +02:00
|
|
|
args = []
|
2018-04-08 15:51:58 -07:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
# do not load .curlrc unless requested (must be the first argument)
|
|
|
|
args << "--disable" unless Homebrew::EnvConfig.curlrc?
|
2018-04-08 15:51:58 -07:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
args << "--globoff"
|
2019-10-01 08:38:44 +02:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
args << "--show-error"
|
2016-06-03 13:05:18 +01:00
|
|
|
|
2020-12-19 17:56:25 -05:00
|
|
|
args << "--user-agent" << case options[:user_agent]
|
2020-10-10 15:23:03 +02:00
|
|
|
when :browser, :fake
|
|
|
|
HOMEBREW_USER_AGENT_FAKE_SAFARI
|
2020-12-19 17:56:25 -05:00
|
|
|
when :default, nil
|
2020-10-10 15:23:03 +02:00
|
|
|
HOMEBREW_USER_AGENT_CURL
|
2020-12-19 17:56:25 -05:00
|
|
|
when String
|
|
|
|
options[:user_agent]
|
2020-10-10 15:23:03 +02:00
|
|
|
end
|
2016-12-25 23:01:40 +00:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
args << "--header" << "Accept-Language: en"
|
2020-09-15 18:51:37 +02:00
|
|
|
|
2020-12-19 17:56:25 -05:00
|
|
|
unless options[:show_output] == true
|
2020-10-10 15:23:03 +02:00
|
|
|
args << "--fail"
|
|
|
|
args << "--progress-bar" unless Context.current.verbose?
|
|
|
|
args << "--verbose" if Homebrew::EnvConfig.curl_verbose?
|
|
|
|
args << "--silent" unless $stdout.tty?
|
|
|
|
end
|
2016-12-25 23:01:40 +00:00
|
|
|
|
2021-03-16 18:15:21 +01:00
|
|
|
args << "--connect-timeout" << connect_timeout.round(3) if options[:connect_timeout]
|
|
|
|
args << "--max-time" << max_time.round(3) if options[:max_time]
|
2020-12-19 17:56:25 -05:00
|
|
|
args << "--retry" << Homebrew::EnvConfig.curl_retries unless options[:retry] == false
|
2021-03-16 18:15:21 +01:00
|
|
|
args << "--retry-max-time" << retry_max_time.round if options[:retry_max_time]
|
2019-05-17 10:14:54 +01:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
args + extra_args
|
2020-09-05 07:41:56 +02:00
|
|
|
end
|
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
def curl_with_workarounds(
|
2021-03-16 18:15:21 +01:00
|
|
|
*args,
|
|
|
|
secrets: nil, print_stdout: nil, print_stderr: nil, debug: nil, verbose: nil, env: {}, timeout: nil, **options
|
2020-10-10 15:23:03 +02:00
|
|
|
)
|
2021-03-16 18:15:21 +01:00
|
|
|
end_time = Time.now + timeout if timeout
|
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
command_options = {
|
|
|
|
secrets: secrets,
|
|
|
|
print_stdout: print_stdout,
|
|
|
|
print_stderr: print_stderr,
|
2020-12-19 11:53:19 -05:00
|
|
|
debug: debug,
|
2020-10-10 15:23:03 +02:00
|
|
|
verbose: verbose,
|
|
|
|
}.compact
|
|
|
|
|
|
|
|
# SSL_CERT_FILE can be incorrectly set by users or portable-ruby and screw
|
|
|
|
# with SSL downloads so unset it here.
|
|
|
|
result = system_command curl_executable,
|
2021-03-16 18:15:21 +01:00
|
|
|
args: curl_args(*args, **options),
|
|
|
|
env: { "SSL_CERT_FILE" => nil }.merge(env),
|
|
|
|
timeout: end_time&.remaining,
|
2020-10-10 15:23:03 +02:00
|
|
|
**command_options
|
|
|
|
|
2021-02-03 09:54:09 +09:00
|
|
|
return result if result.success? || !args.exclude?("--http1.1")
|
2020-10-10 15:23:03 +02:00
|
|
|
|
2021-03-24 01:22:03 -04:00
|
|
|
raise Timeout::Error, result.stderr.chomp if result.status.exitstatus == 28
|
2021-03-16 18:15:21 +01:00
|
|
|
|
2021-02-03 09:54:09 +09:00
|
|
|
# Error in the HTTP2 framing layer
|
2021-03-16 18:15:21 +01:00
|
|
|
if result.status.exitstatus == 16
|
|
|
|
return curl_with_workarounds(
|
|
|
|
*args, "--http1.1",
|
|
|
|
timeout: end_time&.remaining, **command_options, **options
|
|
|
|
)
|
|
|
|
end
|
2020-10-10 15:23:03 +02:00
|
|
|
|
2021-02-03 09:54:09 +09:00
|
|
|
# This is a workaround for https://github.com/curl/curl/issues/1618.
|
|
|
|
if result.status.exitstatus == 56 # Unexpected EOF
|
|
|
|
out = curl_output("-V").stdout
|
2020-10-10 15:23:03 +02:00
|
|
|
|
2021-02-03 09:54:09 +09:00
|
|
|
# If `curl` doesn't support HTTP2, the exception is unrelated to this bug.
|
|
|
|
return result unless out.include?("HTTP2")
|
2020-10-10 15:23:03 +02:00
|
|
|
|
2021-02-03 09:54:09 +09:00
|
|
|
# The bug is fixed in `curl` >= 7.60.0.
|
|
|
|
curl_version = out[/curl (\d+(\.\d+)+)/, 1]
|
|
|
|
return result if Gem::Version.new(curl_version) >= Gem::Version.new("7.60.0")
|
|
|
|
|
|
|
|
return curl_with_workarounds(*args, "--http1.1", **command_options, **options)
|
2020-10-10 15:23:03 +02:00
|
|
|
end
|
|
|
|
|
|
|
|
result
|
2020-09-05 07:41:56 +02:00
|
|
|
end
|
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
def curl(*args, print_stdout: true, **options)
|
|
|
|
result = curl_with_workarounds(*args, print_stdout: print_stdout, **options)
|
|
|
|
result.assert_success!
|
|
|
|
result
|
2020-07-02 18:58:32 +01:00
|
|
|
end
|
2018-09-17 02:45:00 +02:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
def curl_download(*args, to: nil, partial: true, **options)
|
|
|
|
destination = Pathname(to)
|
|
|
|
destination.dirname.mkpath
|
|
|
|
|
|
|
|
if partial
|
|
|
|
range_stdout = curl_output("--location", "--range", "0-1",
|
|
|
|
"--dump-header", "-",
|
|
|
|
"--write-out", "%\{http_code}",
|
|
|
|
"--output", "/dev/null", *args, **options).stdout
|
|
|
|
headers, _, http_status = range_stdout.partition("\r\n\r\n")
|
|
|
|
|
|
|
|
supports_partial_download = http_status.to_i == 206 # Partial Content
|
|
|
|
if supports_partial_download &&
|
|
|
|
destination.exist? &&
|
|
|
|
destination.size == %r{^.*Content-Range: bytes \d+-\d+/(\d+)\r\n.*$}m.match(headers)&.[](1)&.to_i
|
|
|
|
return # We've already downloaded all the bytes
|
|
|
|
end
|
|
|
|
else
|
|
|
|
supports_partial_download = false
|
|
|
|
end
|
|
|
|
|
|
|
|
continue_at = if destination.exist? && supports_partial_download
|
|
|
|
"-"
|
|
|
|
else
|
|
|
|
0
|
|
|
|
end
|
|
|
|
|
|
|
|
curl(
|
|
|
|
"--location", "--remote-time", "--continue-at", continue_at.to_s, "--output", destination, *args, **options
|
|
|
|
)
|
|
|
|
end
|
2017-12-03 14:02:55 +01:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
def curl_output(*args, **options)
|
|
|
|
curl_with_workarounds(*args, print_stderr: false, show_output: true, **options)
|
|
|
|
end
|
2017-12-03 14:02:55 +01:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
# Check if a URL is protected by CloudFlare (e.g. badlion.net and jaxx.io).
|
|
|
|
def url_protected_by_cloudflare?(details)
|
|
|
|
[403, 503].include?(details[:status].to_i) &&
|
|
|
|
details[:headers].match?(/^Set-Cookie: __cfduid=/i) &&
|
|
|
|
details[:headers].match?(/^Server: cloudflare/i)
|
|
|
|
end
|
2017-12-03 14:02:55 +01:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
# Check if a URL is protected by Incapsula (e.g. corsair.com).
|
|
|
|
def url_protected_by_incapsula?(details)
|
|
|
|
details[:status].to_i == 403 &&
|
|
|
|
details[:headers].match?(/^Set-Cookie: visid_incap_/i) &&
|
|
|
|
details[:headers].match?(/^Set-Cookie: incap_ses_/i)
|
|
|
|
end
|
2017-12-03 14:02:55 +01:00
|
|
|
|
2021-01-23 17:26:51 -05:00
|
|
|
def curl_check_http_content(url, specs: {}, user_agents: [:default], check_content: false, strict: false)
|
2020-10-10 15:23:03 +02:00
|
|
|
return unless url.start_with? "http"
|
|
|
|
|
2021-01-02 11:00:15 +01:00
|
|
|
secure_url = url.sub(/\Ahttp:/, "https:")
|
|
|
|
secure_details = nil
|
|
|
|
hash_needed = false
|
|
|
|
if url != secure_url
|
|
|
|
user_agents.each do |user_agent|
|
2021-03-24 01:22:03 -04:00
|
|
|
secure_details =
|
2021-01-23 17:26:51 -05:00
|
|
|
curl_http_content_headers_and_checksum(secure_url, specs: specs, hash_needed: true,
|
|
|
|
user_agent: user_agent)
|
2021-01-02 11:00:15 +01:00
|
|
|
|
|
|
|
next unless http_status_ok?(secure_details[:status])
|
|
|
|
|
|
|
|
hash_needed = true
|
|
|
|
user_agents = [user_agent]
|
|
|
|
break
|
|
|
|
end
|
|
|
|
end
|
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
details = nil
|
2021-01-02 11:00:15 +01:00
|
|
|
user_agents.each do |user_agent|
|
2021-01-23 17:26:51 -05:00
|
|
|
details =
|
|
|
|
curl_http_content_headers_and_checksum(url, specs: specs, hash_needed: hash_needed, user_agent: user_agent)
|
2020-10-10 15:23:03 +02:00
|
|
|
break if http_status_ok?(details[:status])
|
|
|
|
end
|
|
|
|
|
|
|
|
unless details[:status]
|
|
|
|
# Hack around https://github.com/Homebrew/brew/issues/3199
|
|
|
|
return if MacOS.version == :el_capitan
|
|
|
|
|
|
|
|
return "The URL #{url} is not reachable"
|
|
|
|
end
|
|
|
|
|
|
|
|
unless http_status_ok?(details[:status])
|
|
|
|
return if url_protected_by_cloudflare?(details) || url_protected_by_incapsula?(details)
|
|
|
|
|
|
|
|
return "The URL #{url} is not reachable (HTTP status code #{details[:status]})"
|
|
|
|
end
|
|
|
|
|
|
|
|
if url.start_with?("https://") && Homebrew::EnvConfig.no_insecure_redirect? &&
|
|
|
|
!details[:final_url].start_with?("https://")
|
|
|
|
return "The URL #{url} redirects back to HTTP"
|
|
|
|
end
|
|
|
|
|
2021-01-02 11:00:15 +01:00
|
|
|
return unless secure_details
|
2020-10-10 15:23:03 +02:00
|
|
|
|
2021-01-02 11:00:15 +01:00
|
|
|
return if !http_status_ok?(details[:status]) || !http_status_ok?(secure_details[:status])
|
2020-10-10 15:23:03 +02:00
|
|
|
|
|
|
|
etag_match = details[:etag] &&
|
|
|
|
details[:etag] == secure_details[:etag]
|
|
|
|
content_length_match =
|
|
|
|
details[:content_length] &&
|
|
|
|
details[:content_length] == secure_details[:content_length]
|
|
|
|
file_match = details[:file_hash] == secure_details[:file_hash]
|
|
|
|
|
|
|
|
if (etag_match || content_length_match || file_match) &&
|
|
|
|
secure_details[:final_url].start_with?("https://") &&
|
|
|
|
url.start_with?("http://")
|
|
|
|
return "The URL #{url} should use HTTPS rather than HTTP"
|
|
|
|
end
|
|
|
|
|
|
|
|
return unless check_content
|
|
|
|
|
|
|
|
no_protocol_file_contents = %r{https?:\\?/\\?/}
|
2021-01-02 11:00:15 +01:00
|
|
|
http_content = details[:file]&.gsub(no_protocol_file_contents, "/")
|
|
|
|
https_content = secure_details[:file]&.gsub(no_protocol_file_contents, "/")
|
2020-10-10 15:23:03 +02:00
|
|
|
|
|
|
|
# Check for the same content after removing all protocols
|
2021-01-02 11:00:15 +01:00
|
|
|
if (http_content && https_content) && (http_content == https_content) &&
|
|
|
|
url.start_with?("http://") && secure_details[:final_url].start_with?("https://")
|
2020-10-10 15:23:03 +02:00
|
|
|
return "The URL #{url} should use HTTPS rather than HTTP"
|
|
|
|
end
|
|
|
|
|
|
|
|
return unless strict
|
|
|
|
|
|
|
|
# Same size, different content after normalization
|
|
|
|
# (typical causes: Generated ID, Timestamp, Unix time)
|
2021-01-02 11:00:15 +01:00
|
|
|
if http_content.length == https_content.length
|
2020-10-10 15:23:03 +02:00
|
|
|
return "The URL #{url} may be able to use HTTPS rather than HTTP. Please verify it in a browser."
|
|
|
|
end
|
|
|
|
|
2021-01-02 11:00:15 +01:00
|
|
|
lenratio = (100 * https_content.length / http_content.length).to_i
|
2020-10-10 15:23:03 +02:00
|
|
|
return unless (90..110).cover?(lenratio)
|
|
|
|
|
|
|
|
"The URL #{url} may be able to use HTTPS rather than HTTP. Please verify it in a browser."
|
|
|
|
end
|
2018-09-17 02:45:00 +02:00
|
|
|
|
2021-01-23 17:26:51 -05:00
|
|
|
def curl_http_content_headers_and_checksum(url, specs: {}, hash_needed: false, user_agent: :default)
|
2020-10-10 15:23:03 +02:00
|
|
|
file = Tempfile.new.tap(&:close)
|
|
|
|
|
2021-01-23 17:26:51 -05:00
|
|
|
specs = specs.flat_map { |option, argument| ["--#{option.to_s.tr("_", "-")}", argument] }
|
2020-10-10 15:23:03 +02:00
|
|
|
max_time = hash_needed ? "600" : "25"
|
2021-01-02 11:00:15 +01:00
|
|
|
output, _, status = curl_output(
|
2021-01-23 17:26:51 -05:00
|
|
|
*specs, "--dump-header", "-", "--output", file.path, "--location",
|
2021-01-02 11:00:15 +01:00
|
|
|
"--connect-timeout", "15", "--max-time", max_time, "--retry-max-time", max_time, url,
|
2020-10-10 15:23:03 +02:00
|
|
|
user_agent: user_agent
|
|
|
|
)
|
|
|
|
|
|
|
|
status_code = :unknown
|
|
|
|
while status_code == :unknown || status_code.to_s.start_with?("3")
|
|
|
|
headers, _, output = output.partition("\r\n\r\n")
|
|
|
|
status_code = headers[%r{HTTP/.* (\d+)}, 1]
|
|
|
|
location = headers[/^Location:\s*(.*)$/i, 1]
|
|
|
|
final_url = location.chomp if location
|
|
|
|
end
|
|
|
|
|
2021-01-02 11:00:15 +01:00
|
|
|
if status.success?
|
|
|
|
file_contents = File.read(file.path)
|
|
|
|
file_hash = Digest::SHA2.hexdigest(file_contents) if hash_needed
|
|
|
|
end
|
2020-10-10 15:23:03 +02:00
|
|
|
|
|
|
|
final_url ||= url
|
|
|
|
|
|
|
|
{
|
|
|
|
url: url,
|
|
|
|
final_url: final_url,
|
|
|
|
status: status_code,
|
|
|
|
etag: headers[%r{ETag: ([wW]/)?"(([^"]|\\")*)"}, 2],
|
|
|
|
content_length: headers[/Content-Length: (\d+)/, 1],
|
|
|
|
headers: headers,
|
2020-12-10 07:35:43 +01:00
|
|
|
file_hash: file_hash,
|
2021-01-02 11:00:15 +01:00
|
|
|
file: file_contents,
|
2020-10-10 15:23:03 +02:00
|
|
|
}
|
|
|
|
ensure
|
|
|
|
file.unlink
|
|
|
|
end
|
2017-12-03 14:02:55 +01:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
def http_status_ok?(status)
|
|
|
|
(100..299).cover?(status.to_i)
|
|
|
|
end
|
2017-12-03 14:02:55 +01:00
|
|
|
end
|
|
|
|
end
|
2019-09-18 10:32:13 +01:00
|
|
|
|
2020-10-10 15:23:03 +02:00
|
|
|
# FIXME: Include `Utils::Curl` explicitly everywhere it is used.
|
|
|
|
include Utils::Curl # rubocop:disable Style/MixinUsage
|