require 'utils/json' require 'erb' class AbstractDownloadStrategy attr_reader :name, :resource def initialize name, resource @name = name @resource = resource @url = resource.url end def expand_safe_system_args args args = args.dup args.each_with_index do |arg, ii| if arg.is_a? Hash unless ARGV.verbose? args[ii] = arg[:quiet_flag] else args.delete_at ii end return args end end # 2 as default because commands are eg. svn up, git pull args.insert(2, '-q') unless ARGV.verbose? args end def quiet_safe_system *args safe_system(*expand_safe_system_args(args)) end # All download strategies are expected to implement these methods def fetch; end def stage; end def cached_location; end def clear_cache; end end class VCSDownloadStrategy < AbstractDownloadStrategy REF_TYPES = [:branch, :revision, :revisions, :tag].freeze def initialize name, resource super @ref_type, @ref = extract_ref(resource.specs) @clone = HOMEBREW_CACHE/cache_filename end def extract_ref(specs) key = REF_TYPES.find { |type| specs.key?(type) } return key, specs[key] end def cache_filename(tag=cache_tag) if name.empty? || name == '__UNKNOWN__' "#{ERB::Util.url_encode(@url)}--#{tag}" else "#{name}--#{tag}" end end def cache_tag "__UNKNOWN__" end def cached_location @clone end def clear_cache cached_location.rmtree if cached_location.exist? end end class CurlDownloadStrategy < AbstractDownloadStrategy def mirrors @mirrors ||= resource.mirrors.dup end def tarball_path @tarball_path ||= if name.empty? || name == '__UNKNOWN__' Pathname.new("#{HOMEBREW_CACHE}/#{basename_without_params}") else Pathname.new("#{HOMEBREW_CACHE}/#{name}-#{resource.version}#{ext}") end end def temporary_path @temporary_path ||= Pathname.new("#{tarball_path}.incomplete") end def cached_location tarball_path end def clear_cache [cached_location, temporary_path].each { |f| f.unlink if f.exist? } end def downloaded_size temporary_path.size? or 0 end # Private method, can be overridden if needed. def _fetch curl @url, '-C', downloaded_size, '-o', temporary_path end def fetch ohai "Downloading #{@url}" unless tarball_path.exist? had_incomplete_download = temporary_path.exist? begin _fetch rescue ErrorDuringExecution # 33 == range not supported # try wiping the incomplete download and retrying once if $?.exitstatus == 33 && had_incomplete_download ohai "Trying a full download" temporary_path.unlink had_incomplete_download = false retry else raise CurlDownloadStrategyError, "Download failed: #{@url}" end end ignore_interrupts { temporary_path.rename(tarball_path) } else puts "Already downloaded: #{tarball_path}" end rescue CurlDownloadStrategyError raise if mirrors.empty? puts "Trying a mirror..." @url = mirrors.shift retry else tarball_path end # gunzip and bunzip2 write the output file in the same directory as the input # file regardless of the current working directory, so we need to write it to # the correct location ourselves. def buffered_write(tool) target = File.basename(basename_without_params, tarball_path.extname) IO.popen("#{tool} -f '#{tarball_path}' -c") do |pipe| File.open(target, "wb") do |f| buf = "" f.write(buf) while pipe.read(1024, buf) end end end def stage case tarball_path.compression_type when :zip with_system_path { quiet_safe_system 'unzip', {:quiet_flag => '-qq'}, tarball_path } chdir when :gzip_only with_system_path { buffered_write("gunzip") } when :bzip2_only with_system_path { buffered_write("bunzip2") } when :gzip, :bzip2, :compress, :tar # Assume these are also tarred # TODO check if it's really a tar archive with_system_path { safe_system 'tar', 'xf', tarball_path } chdir when :xz with_system_path { safe_system "#{xzpath} -dc \"#{tarball_path}\" | tar xf -" } chdir when :lzip with_system_path { safe_system "#{lzippath} -dc \"#{tarball_path}\" | tar xf -" } chdir when :xar safe_system "/usr/bin/xar", "-xf", tarball_path when :rar quiet_safe_system 'unrar', 'x', {:quiet_flag => '-inul'}, tarball_path when :p7zip safe_system '7zr', 'x', tarball_path else FileUtils.cp tarball_path, basename_without_params end end private def curl(*args) args << '--connect-timeout' << '5' unless mirrors.empty? super end def xzpath "#{HOMEBREW_PREFIX}/opt/xz/bin/xz" end def lzippath "#{HOMEBREW_PREFIX}/opt/lzip/bin/lzip" end def chdir entries = Dir['*'] case entries.length when 0 then raise "Empty archive" when 1 then Dir.chdir entries.first rescue nil end end def basename_without_params # Strip any ?thing=wad out of .c?thing=wad style extensions File.basename(@url)[/[^?]+/] end def ext # GitHub uses odd URLs for zip files, so check for those rx=%r[https?://(www\.)?github\.com/.*/(zip|tar)ball/] if rx.match @url if $2 == 'zip' '.zip' else '.tgz' end else # Strip any ?thing=wad out of .c?thing=wad style extensions (Pathname.new(@url).extname)[/[^?]+/] end end end # Detect and download from Apache Mirror class CurlApacheMirrorDownloadStrategy < CurlDownloadStrategy def apache_mirrors rd, wr = IO.pipe buf = "" pid = fork do rd.close $stdout.reopen(wr) $stderr.reopen(wr) curl "#{@url}&asjson=1" end wr.close rd.readline if ARGV.verbose? # Remove Homebrew output buf << rd.read until rd.eof? rd.close Process.wait(pid) buf end def _fetch return super if @tried_apache_mirror @tried_apache_mirror = true mirrors = Utils::JSON.load(apache_mirrors) @url = mirrors.fetch('preferred') + mirrors.fetch('path_info') ohai "Best Mirror #{@url}" super rescue IndexError, Utils::JSON::Error raise CurlDownloadStrategyError, "Couldn't determine mirror, try again later." end end # Download via an HTTP POST. # Query parameters on the URL are converted into POST parameters class CurlPostDownloadStrategy < CurlDownloadStrategy def _fetch base_url,data = @url.split('?') curl base_url, '-d', data, '-C', downloaded_size, '-o', temporary_path end end # Download from an SSL3-only host. class CurlSSL3DownloadStrategy < CurlDownloadStrategy def _fetch curl @url, '-3', '-C', downloaded_size, '-o', temporary_path end end # Use this strategy to download but not unzip a file. # Useful for installing jars. class NoUnzipCurlDownloadStrategy < CurlDownloadStrategy def stage FileUtils.cp tarball_path, basename_without_params end end # This strategy is provided for use with sites that only provide HTTPS and # also have a broken cert. Try not to need this, as we probably won't accept # the formula. class CurlUnsafeDownloadStrategy < CurlDownloadStrategy def _fetch curl @url, '--insecure', '-C', downloaded_size, '-o', temporary_path end end # This strategy extracts our binary packages. class CurlBottleDownloadStrategy < CurlDownloadStrategy def initialize name, resource super mirror = ENV['HOMEBREW_SOURCEFORGE_MIRROR'] @url = "#{@url}?use_mirror=#{mirror}" if mirror end def tarball_path @tarball_path ||= HOMEBREW_CACHE/"#{name}-#{resource.version}#{ext}" end def stage ohai "Pouring #{tarball_path.basename}" super end end # This strategy extracts local binary packages. class LocalBottleDownloadStrategy < CurlDownloadStrategy def initialize formula super formula.name, formula.active_spec @tarball_path = formula.local_bottle_path end def stage ohai "Pouring #{tarball_path.basename}" super end end # S3DownloadStrategy downloads tarballs from AWS S3. # To use it, add ":using => S3DownloadStrategy" to the URL section of your # formula. This download strategy uses AWS access tokens (in the # environment variables AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY) # to sign the request. This strategy is good in a corporate setting, # because it lets you use a private S3 bucket as a repo for internal # distribution. (It will work for public buckets as well.) class S3DownloadStrategy < CurlDownloadStrategy def _fetch # Put the aws gem requirement here (vs top of file) so it's only # a dependency of S3 users, not all Homebrew users require 'rubygems' begin require 'aws-sdk' rescue LoadError onoe "Install the aws-sdk gem into the gem repo used by brew." raise end if @url !~ %r[^https?://+([^.]+).s3.amazonaws.com/+(.+)$] then raise "Bad S3 URL: " + @url end (bucket,key) = $1,$2 obj = AWS::S3.new().buckets[bucket].objects[key] begin s3url = obj.url_for(:get) rescue AWS::Errors::MissingCredentialsError ohai "AWS credentials missing, trying public URL instead." s3url = obj.public_url end curl s3url, '-C', downloaded_size, '-o', temporary_path end end class SubversionDownloadStrategy < VCSDownloadStrategy def cache_tag ARGV.build_head? ? "svn-HEAD" : "svn" end def repo_valid? @clone.join(".svn").directory? end def repo_url `svn info '#{@clone}' 2>/dev/null`.strip[/^URL: (.+)$/, 1] end def fetch @url = @url.sub(/^svn\+/, '') if @url =~ %r[^svn\+http://] ohai "Checking out #{@url}" clear_cache unless @url.chomp("/") == repo_url or quiet_system 'svn', 'switch', @url, @clone if @clone.exist? and not repo_valid? puts "Removing invalid SVN repo from cache" clear_cache end case @ref_type when :revision fetch_repo @clone, @url, @ref when :revisions # nil is OK for main_revision, as fetch_repo will then get latest main_revision = @ref[:trunk] fetch_repo @clone, @url, main_revision, true get_externals do |external_name, external_url| fetch_repo @clone+external_name, external_url, @ref[external_name], true end else fetch_repo @clone, @url end end def stage quiet_safe_system 'svn', 'export', '--force', @clone, Dir.pwd end def shell_quote str # Oh god escaping shell args. # See http://notetoself.vrensk.com/2008/08/escaping-single-quotes-in-ruby-harder-than-expected/ str.gsub(/\\|'/) { |c| "\\#{c}" } end def get_externals `svn propget svn:externals '#{shell_quote(@url)}'`.chomp.each_line do |line| name, url = line.split(/\s+/) yield name, url end end def fetch_repo target, url, revision=nil, ignore_externals=false # Use "svn up" when the repository already exists locally. # This saves on bandwidth and will have a similar effect to verifying the # cache as it will make any changes to get the right revision. svncommand = target.directory? ? 'up' : 'checkout' args = ['svn', svncommand] # SVN shipped with XCode 3.1.4 can't force a checkout. args << '--force' unless MacOS.version == :leopard args << url unless target.directory? args << target args << '-r' << revision if revision args << '--ignore-externals' if ignore_externals quiet_safe_system(*args) end end # Require a newer version of Subversion than 1.4.x (Leopard-provided version) class StrictSubversionDownloadStrategy < SubversionDownloadStrategy def find_svn exe = `svn -print-path` `#{exe} --version` =~ /version (\d+\.\d+(\.\d+)*)/ svn_version = $1 version_tuple=svn_version.split(".").collect {|v|Integer(v)} if version_tuple[0] == 1 and version_tuple[1] <= 4 onoe "Detected Subversion (#{exe}, version #{svn_version}) is too old." puts "Subversion 1.4.x will not export externals correctly for this formula." puts "You must either `brew install subversion` or set HOMEBREW_SVN to the path" puts "of a newer svn binary." end return exe end end # Download from SVN servers with invalid or self-signed certs class UnsafeSubversionDownloadStrategy < SubversionDownloadStrategy def fetch_repo target, url, revision=nil, ignore_externals=false # Use "svn up" when the repository already exists locally. # This saves on bandwidth and will have a similar effect to verifying the # cache as it will make any changes to get the right revision. svncommand = target.directory? ? 'up' : 'checkout' args = ['svn', svncommand, '--non-interactive', '--trust-server-cert', '--force'] args << url unless target.directory? args << target args << '-r' << revision if revision args << '--ignore-externals' if ignore_externals quiet_safe_system(*args) end end class GitDownloadStrategy < VCSDownloadStrategy SHALLOW_CLONE_WHITELIST = [ %r{git://}, %r{https://github\.com}, %r{http://git\.sv\.gnu\.org}, %r{http://llvm\.org}, ] def initialize name, resource super @shallow = resource.specs.fetch(:shallow) { true } end def cache_tag; "git" end def fetch ohai "Cloning #@url" if @clone.exist? && repo_valid? puts "Updating #@clone" @clone.cd do config_repo update_repo checkout reset update_submodules if submodules? end elsif @clone.exist? puts "Removing invalid .git repo from cache" clear_cache clone_repo else clone_repo end end def stage dst = Dir.getwd @clone.cd do if @ref_type and @ref ohai "Checking out #@ref_type #@ref" else reset end # http://stackoverflow.com/questions/160608/how-to-do-a-git-export-like-svn-export safe_system 'git', 'checkout-index', '-a', '-f', "--prefix=#{dst}/" checkout_submodules(dst) if submodules? end end private def shallow_clone? @shallow && support_depth? end def support_depth? @ref_type != :revision && SHALLOW_CLONE_WHITELIST.any? { |rx| rx === @url } end def git_dir @clone.join(".git") end def has_ref? quiet_system 'git', '--git-dir', git_dir, 'rev-parse', '-q', '--verify', @ref end def repo_valid? quiet_system "git", "--git-dir", git_dir, "status", "-s" end def submodules? @clone.join(".gitmodules").exist? end def clone_args args = %w{clone} args << '--depth' << '1' if shallow_clone? case @ref_type when :branch, :tag then args << '--branch' << @ref end args << @url << @clone end def refspec case @ref_type when :branch then "+refs/heads/#@ref:refs/remotes/origin/#@ref" when :tag then "+refs/tags/#@ref:refs/tags/#@ref" else "+refs/heads/master:refs/remotes/origin/master" end end def config_repo safe_system 'git', 'config', 'remote.origin.url', @url safe_system 'git', 'config', 'remote.origin.fetch', refspec end def update_repo # Branches always need updated. The has_ref? check will only work if a ref # has been specified; if there isn't one we always want an update. if @ref_type == :branch || !@ref || !has_ref? quiet_safe_system 'git', 'fetch', 'origin' end end def clone_repo safe_system 'git', *clone_args @clone.cd { update_submodules } if submodules? end def checkout_args ref = case @ref_type when :branch, :tag, :revision then @ref else `git symbolic-ref refs/remotes/origin/HEAD`.strip.split("/").last end %W{checkout -f #{ref}} end def checkout quiet_safe_system 'git', *checkout_args end def reset_args ref = case @ref_type when :branch then "origin/#@ref" when :revision, :tag then @ref else "origin/HEAD" end %W{reset --hard #{ref}} end def reset quiet_safe_system 'git', *reset_args end def update_submodules safe_system 'git', 'submodule', 'update', '--init' end def checkout_submodules(dst) sub_cmd = "git checkout-index -a -f --prefix=#{dst}/$path/" safe_system 'git', 'submodule', '--quiet', 'foreach', '--recursive', sub_cmd end end class CVSDownloadStrategy < VCSDownloadStrategy def cvspath @path ||= %W[ /usr/bin/cvs #{HOMEBREW_PREFIX}/bin/cvs #{HOMEBREW_PREFIX}/opt/cvs/bin/cvs #{which("cvs")} ].find { |p| File.executable? p } end def cache_tag; "cvs" end def fetch ohai "Checking out #{@url}" # URL of cvs cvs://:pserver:anoncvs@www.gccxml.org:/cvsroot/GCC_XML:gccxml # will become: # cvs -d :pserver:anoncvs@www.gccxml.org:/cvsroot/GCC_XML login # cvs -d :pserver:anoncvs@www.gccxml.org:/cvsroot/GCC_XML co gccxml mod, url = split_url(@url) unless @clone.exist? HOMEBREW_CACHE.cd do safe_system cvspath, '-d', url, 'login' safe_system cvspath, '-d', url, 'checkout', '-d', cache_filename("cvs"), mod end else puts "Updating #{@clone}" @clone.cd { safe_system cvspath, 'up' } end end def stage FileUtils.cp_r Dir[@clone+"{.}"], Dir.pwd require 'find' Find.find(Dir.pwd) do |path| if FileTest.directory?(path) && File.basename(path) == "CVS" Find.prune FileUtil.rm_r path, :force => true end end end private def split_url(in_url) parts=in_url.sub(%r[^cvs://], '').split(/:/) mod=parts.pop url=parts.join(':') [ mod, url ] end end class MercurialDownloadStrategy < VCSDownloadStrategy def cache_tag; "hg" end def hgpath # Note: #{HOMEBREW_PREFIX}/share/python/hg is deprecated @path ||= %W[ #{which("hg")} #{HOMEBREW_PREFIX}/bin/hg #{HOMEBREW_PREFIX}/opt/mercurial/bin/hg #{HOMEBREW_PREFIX}/share/python/hg ].find { |p| File.executable? p } end def fetch ohai "Cloning #{@url}" if @clone.exist? && repo_valid? puts "Updating #{@clone}" @clone.cd { quiet_safe_system hgpath, 'pull', '--update' } elsif @clone.exist? puts "Removing invalid hg repo from cache" clear_cache clone_repo else clone_repo end end def repo_valid? @clone.join(".hg").directory? end def clone_repo url = @url.sub(%r[^hg://], '') safe_system hgpath, 'clone', url, @clone end def stage dst = Dir.getwd @clone.cd do if @ref_type and @ref ohai "Checking out #{@ref_type} #{@ref}" safe_system hgpath, 'archive', '--subrepos', '-y', '-r', @ref, '-t', 'files', dst else safe_system hgpath, 'archive', '--subrepos', '-y', '-t', 'files', dst end end end end class BazaarDownloadStrategy < VCSDownloadStrategy def cache_tag; "bzr" end def bzrpath @path ||= %W[ #{which("bzr")} #{HOMEBREW_PREFIX}/bin/bzr ].find { |p| File.executable? p } end def repo_valid? @clone.join(".bzr").directory? end def fetch ohai "Cloning #{@url}" if @clone.exist? && repo_valid? puts "Updating #{@clone}" @clone.cd { safe_system bzrpath, 'update' } elsif @clone.exist? puts "Removing invalid bzr repo from cache" clear_cache clone_repo else clone_repo end end def clone_repo url = @url.sub(%r[^bzr://], '') # 'lightweight' means history-less safe_system bzrpath, 'checkout', '--lightweight', url, @clone end def stage # FIXME: The export command doesn't work on checkouts # See https://bugs.launchpad.net/bzr/+bug/897511 FileUtils.cp_r Dir[@clone+"{.}"], Dir.pwd FileUtils.rm_r Dir[Dir.pwd+"/.bzr"] end end class FossilDownloadStrategy < VCSDownloadStrategy def cache_tag; "fossil" end def fossilpath @path ||= %W[ #{which("fossil")} #{HOMEBREW_PREFIX}/bin/fossil ].find { |p| File.executable? p } end def fetch ohai "Cloning #{@url}" unless @clone.exist? url=@url.sub(%r[^fossil://], '') safe_system fossilpath, 'clone', url, @clone else puts "Updating #{@clone}" safe_system fossilpath, 'pull', '-R', @clone end end def stage # TODO: The 'open' and 'checkout' commands are very noisy and have no '-q' option. safe_system fossilpath, 'open', @clone if @ref_type and @ref ohai "Checking out #{@ref_type} #{@ref}" safe_system fossilpath, 'checkout', @ref end end end class DownloadStrategyDetector def self.detect(url, strategy=nil) if strategy.nil? detect_from_url(url) elsif Class === strategy && strategy < AbstractDownloadStrategy strategy elsif Symbol === strategy detect_from_symbol(strategy) else raise TypeError, "Unknown download strategy specification #{strategy.inspect}" end end def self.detect_from_url(url) case url when %r[^https?://.+\.git$], %r[^git://] GitDownloadStrategy when %r[^http://www\.apache\.org/dyn/closer\.cgi] CurlApacheMirrorDownloadStrategy when %r[^https?://(.+?\.)?googlecode\.com/svn], %r[^https?://svn\.], %r[^svn://], %r[^https?://(.+?\.)?sourceforge\.net/svnroot/] SubversionDownloadStrategy when %r[^cvs://] CVSDownloadStrategy when %r[^https?://(.+?\.)?googlecode\.com/hg] MercurialDownloadStrategy when %r[^hg://] MercurialDownloadStrategy when %r[^bzr://] BazaarDownloadStrategy when %r[^fossil://] FossilDownloadStrategy when %r[^http://svn\.apache\.org/repos/], %r[^svn\+http://] SubversionDownloadStrategy when %r[^https?://(.+?\.)?sourceforge\.net/hgweb/] MercurialDownloadStrategy else CurlDownloadStrategy end end def self.detect_from_symbol(symbol) case symbol when :hg then MercurialDownloadStrategy when :nounzip then NoUnzipCurlDownloadStrategy when :git then GitDownloadStrategy when :bzr then BazaarDownloadStrategy when :svn then SubversionDownloadStrategy when :curl then CurlDownloadStrategy when :ssl3 then CurlSSL3DownloadStrategy when :cvs then CVSDownloadStrategy when :post then CurlPostDownloadStrategy else raise "Unknown download strategy #{strategy} was requested." end end end