384 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			Ruby
		
	
	
	
	
	
			
		
		
	
	
			384 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			Ruby
		
	
	
	
	
	
# frozen_string_literal: true
 | 
						|
 | 
						|
require_relative '../../config/boot'
 | 
						|
require_relative '../../config/environment'
 | 
						|
require_relative 'cli_helper'
 | 
						|
 | 
						|
module Mastodon
 | 
						|
  class MediaCLI < Thor
 | 
						|
    include ActionView::Helpers::NumberHelper
 | 
						|
    include CLIHelper
 | 
						|
 | 
						|
    def self.exit_on_failure?
 | 
						|
      true
 | 
						|
    end
 | 
						|
 | 
						|
    option :days, type: :numeric, default: 7, aliases: [:d]
 | 
						|
    option :prune_profiles, type: :boolean, default: false
 | 
						|
    option :remove_headers, type: :boolean, default: false
 | 
						|
    option :include_follows, type: :boolean, default: false
 | 
						|
    option :concurrency, type: :numeric, default: 5, aliases: [:c]
 | 
						|
    option :dry_run, type: :boolean, default: false
 | 
						|
    desc 'remove', 'Remove remote media files, headers or avatars'
 | 
						|
    long_desc <<-DESC
 | 
						|
      Removes locally cached copies of media attachments (and optionally profile
 | 
						|
      headers and avatars) from other servers. By default, only media attachements
 | 
						|
      are removed.
 | 
						|
      The --days option specifies how old media attachments have to be before
 | 
						|
      they are removed. In case of avatars and headers, it specifies how old
 | 
						|
      the last webfinger request and update to the user has to be before they
 | 
						|
      are pruned. It defaults to 7 days.
 | 
						|
      If --prune-profiles is specified, only avatars and headers are removed.
 | 
						|
      If --remove-headers is specified, only headers are removed.
 | 
						|
      If --include-follows is specified along with --prune-profiles or
 | 
						|
      --remove-headers, all non-local profiles will be pruned irrespective of
 | 
						|
      follow status. By default, only accounts that are not followed by or
 | 
						|
      following anyone locally are pruned.
 | 
						|
    DESC
 | 
						|
    # rubocop:disable Metrics/PerceivedComplexity
 | 
						|
    def remove
 | 
						|
      if options[:prune_profiles] && options[:remove_headers]
 | 
						|
        say('--prune-profiles and --remove-headers should not be specified simultaneously', :red, true)
 | 
						|
        exit(1)
 | 
						|
      end
 | 
						|
      if options[:include_follows] && !(options[:prune_profiles] || options[:remove_headers])
 | 
						|
        say('--include-follows can only be used with --prune-profiles or --remove-headers', :red, true)
 | 
						|
        exit(1)
 | 
						|
      end
 | 
						|
      time_ago        = options[:days].days.ago
 | 
						|
      dry_run         = options[:dry_run] ? ' (DRY RUN)' : ''
 | 
						|
 | 
						|
      if options[:prune_profiles] || options[:remove_headers]
 | 
						|
        processed, aggregate = parallelize_with_progress(Account.remote.where({ last_webfingered_at: ..time_ago, updated_at: ..time_ago })) do |account|
 | 
						|
          next if !options[:include_follows] && Follow.where(account: account).or(Follow.where(target_account: account)).exists?
 | 
						|
          next if account.avatar.blank? && account.header.blank?
 | 
						|
          next if options[:remove_headers] && account.header.blank?
 | 
						|
 | 
						|
          size = (account.header_file_size || 0)
 | 
						|
          size += (account.avatar_file_size || 0) if options[:prune_profiles]
 | 
						|
 | 
						|
          unless options[:dry_run]
 | 
						|
            account.header.destroy
 | 
						|
            account.avatar.destroy if options[:prune_profiles]
 | 
						|
            account.save!
 | 
						|
          end
 | 
						|
 | 
						|
          size
 | 
						|
        end
 | 
						|
 | 
						|
        say("Visited #{processed} accounts and removed profile media totaling #{number_to_human_size(aggregate)}#{dry_run}", :green, true)
 | 
						|
      end
 | 
						|
 | 
						|
      unless options[:prune_profiles] || options[:remove_headers]
 | 
						|
        processed, aggregate = parallelize_with_progress(MediaAttachment.cached.where.not(remote_url: '').where(created_at: ..time_ago)) do |media_attachment|
 | 
						|
          next if media_attachment.file.blank?
 | 
						|
 | 
						|
          size = (media_attachment.file_file_size || 0) + (media_attachment.thumbnail_file_size || 0)
 | 
						|
 | 
						|
          unless options[:dry_run]
 | 
						|
            media_attachment.file.destroy
 | 
						|
            media_attachment.thumbnail.destroy
 | 
						|
            media_attachment.save
 | 
						|
          end
 | 
						|
 | 
						|
          size
 | 
						|
        end
 | 
						|
 | 
						|
        say("Removed #{processed} media attachments (approx. #{number_to_human_size(aggregate)})#{dry_run}", :green, true)
 | 
						|
      end
 | 
						|
    end
 | 
						|
 | 
						|
    option :start_after
 | 
						|
    option :prefix
 | 
						|
    option :fix_permissions, type: :boolean, default: false
 | 
						|
    option :dry_run, type: :boolean, default: false
 | 
						|
    desc 'remove-orphans', 'Scan storage and check for files that do not belong to existing media attachments'
 | 
						|
    long_desc <<~LONG_DESC
 | 
						|
      Scans file storage for files that do not belong to existing media attachments. Because this operation
 | 
						|
      requires iterating over every single file individually, it will be slow.
 | 
						|
 | 
						|
      Please mind that some storage providers charge for the necessary API requests to list objects.
 | 
						|
    LONG_DESC
 | 
						|
    def remove_orphans
 | 
						|
      progress        = create_progress_bar(nil)
 | 
						|
      reclaimed_bytes = 0
 | 
						|
      removed         = 0
 | 
						|
      dry_run         = options[:dry_run] ? ' (DRY RUN)' : ''
 | 
						|
      prefix          = options[:prefix]
 | 
						|
 | 
						|
      case Paperclip::Attachment.default_options[:storage]
 | 
						|
      when :s3
 | 
						|
        paperclip_instance = MediaAttachment.new.file
 | 
						|
        s3_interface       = paperclip_instance.s3_interface
 | 
						|
        s3_permissions     = Paperclip::Attachment.default_options[:s3_permissions]
 | 
						|
        bucket             = s3_interface.bucket(Paperclip::Attachment.default_options[:s3_credentials][:bucket])
 | 
						|
        last_key           = options[:start_after]
 | 
						|
 | 
						|
        loop do
 | 
						|
          objects = begin
 | 
						|
            bucket.objects(start_after: last_key, prefix: prefix).limit(1000).map { |x| x }
 | 
						|
          rescue => e
 | 
						|
            progress.log(pastel.red("Error fetching list of files: #{e}"))
 | 
						|
            progress.log("If you want to continue from this point, add --start-after=#{last_key} to your command") if last_key
 | 
						|
            break
 | 
						|
          end
 | 
						|
 | 
						|
          break if objects.empty?
 | 
						|
 | 
						|
          last_key   = objects.last.key
 | 
						|
          record_map = preload_records_from_mixed_objects(objects)
 | 
						|
 | 
						|
          objects.each do |object|
 | 
						|
            object.acl.put(acl: s3_permissions) if options[:fix_permissions] && !options[:dry_run]
 | 
						|
 | 
						|
            path_segments = object.key.split('/')
 | 
						|
            path_segments.delete('cache')
 | 
						|
 | 
						|
            unless [7, 10].include?(path_segments.size)
 | 
						|
              progress.log(pastel.yellow("Unrecognized file found: #{object.key}"))
 | 
						|
              next
 | 
						|
            end
 | 
						|
 | 
						|
            model_name      = path_segments.first.classify
 | 
						|
            attachment_name = path_segments[1].singularize
 | 
						|
            record_id       = path_segments[2..-2].join.to_i
 | 
						|
            file_name       = path_segments.last
 | 
						|
            record          = record_map.dig(model_name, record_id)
 | 
						|
            attachment      = record&.public_send(attachment_name)
 | 
						|
 | 
						|
            progress.increment
 | 
						|
 | 
						|
            next unless attachment.blank? || !attachment.variant?(file_name)
 | 
						|
 | 
						|
            begin
 | 
						|
              object.delete unless options[:dry_run]
 | 
						|
 | 
						|
              reclaimed_bytes += object.size
 | 
						|
              removed += 1
 | 
						|
 | 
						|
              progress.log("Found and removed orphan: #{object.key}")
 | 
						|
            rescue => e
 | 
						|
              progress.log(pastel.red("Error processing #{object.key}: #{e}"))
 | 
						|
            end
 | 
						|
          end
 | 
						|
        end
 | 
						|
      when :fog
 | 
						|
        say('The fog storage driver is not supported for this operation at this time', :red)
 | 
						|
        exit(1)
 | 
						|
      when :filesystem
 | 
						|
        require 'find'
 | 
						|
 | 
						|
        root_path = ENV.fetch('PAPERCLIP_ROOT_PATH', File.join(':rails_root', 'public', 'system')).gsub(':rails_root', Rails.root.to_s)
 | 
						|
 | 
						|
        Find.find(File.join(*[root_path, prefix].compact)) do |path|
 | 
						|
          next if File.directory?(path)
 | 
						|
 | 
						|
          key = path.gsub("#{root_path}#{File::SEPARATOR}", '')
 | 
						|
 | 
						|
          path_segments = key.split(File::SEPARATOR)
 | 
						|
          path_segments.delete('cache')
 | 
						|
 | 
						|
          unless [7, 10].include?(path_segments.size)
 | 
						|
            progress.log(pastel.yellow("Unrecognized file found: #{key}"))
 | 
						|
            next
 | 
						|
          end
 | 
						|
 | 
						|
          model_name      = path_segments.first.classify
 | 
						|
          record_id       = path_segments[2..-2].join.to_i
 | 
						|
          attachment_name = path_segments[1].singularize
 | 
						|
          file_name       = path_segments.last
 | 
						|
 | 
						|
          next unless PRELOAD_MODEL_WHITELIST.include?(model_name)
 | 
						|
 | 
						|
          record     = model_name.constantize.find_by(id: record_id)
 | 
						|
          attachment = record&.public_send(attachment_name)
 | 
						|
 | 
						|
          progress.increment
 | 
						|
 | 
						|
          next unless attachment.blank? || !attachment.variant?(file_name)
 | 
						|
 | 
						|
          begin
 | 
						|
            size = File.size(path)
 | 
						|
 | 
						|
            unless options[:dry_run]
 | 
						|
              File.delete(path)
 | 
						|
              begin
 | 
						|
                FileUtils.rmdir(File.dirname(path), parents: true)
 | 
						|
              rescue Errno::ENOTEMPTY
 | 
						|
                # OK
 | 
						|
              end
 | 
						|
            end
 | 
						|
 | 
						|
            reclaimed_bytes += size
 | 
						|
            removed += 1
 | 
						|
 | 
						|
            progress.log("Found and removed orphan: #{key}")
 | 
						|
          rescue => e
 | 
						|
            progress.log(pastel.red("Error processing #{key}: #{e}"))
 | 
						|
          end
 | 
						|
        end
 | 
						|
      end
 | 
						|
 | 
						|
      progress.total = progress.progress
 | 
						|
      progress.finish
 | 
						|
 | 
						|
      say("Removed #{removed} orphans (approx. #{number_to_human_size(reclaimed_bytes)})#{dry_run}", :green, true)
 | 
						|
    end
 | 
						|
    # rubocop:enable Metrics/PerceivedComplexity
 | 
						|
 | 
						|
    option :account, type: :string
 | 
						|
    option :domain, type: :string
 | 
						|
    option :status, type: :numeric
 | 
						|
    option :days, type: :numeric
 | 
						|
    option :concurrency, type: :numeric, default: 5, aliases: [:c]
 | 
						|
    option :verbose, type: :boolean, default: false, aliases: [:v]
 | 
						|
    option :dry_run, type: :boolean, default: false
 | 
						|
    option :force, type: :boolean, default: false
 | 
						|
    desc 'refresh', 'Fetch remote media files'
 | 
						|
    long_desc <<-DESC
 | 
						|
      Re-downloads media attachments from other servers. You must specify the
 | 
						|
      source of media attachments with one of the following options:
 | 
						|
 | 
						|
      Use the --status option to download attachments from a specific status,
 | 
						|
      using the status local numeric ID.
 | 
						|
 | 
						|
      Use the --account option to download attachments from a specific account,
 | 
						|
      using username@domain handle of the account.
 | 
						|
 | 
						|
      Use the --domain option to download attachments from a specific domain.
 | 
						|
 | 
						|
      Use the --days option to limit attachments created within days.
 | 
						|
 | 
						|
      By default, attachments that are believed to be already downloaded will
 | 
						|
      not be re-downloaded. To force re-download of every URL, use --force.
 | 
						|
    DESC
 | 
						|
    def refresh
 | 
						|
      dry_run = options[:dry_run] ? ' (DRY RUN)' : ''
 | 
						|
 | 
						|
      if options[:status]
 | 
						|
        scope = MediaAttachment.where(status_id: options[:status])
 | 
						|
      elsif options[:account]
 | 
						|
        username, domain = options[:account].split('@')
 | 
						|
        account = Account.find_remote(username, domain)
 | 
						|
 | 
						|
        if account.nil?
 | 
						|
          say('No such account', :red)
 | 
						|
          exit(1)
 | 
						|
        end
 | 
						|
 | 
						|
        scope = MediaAttachment.where(account_id: account.id)
 | 
						|
      elsif options[:domain]
 | 
						|
        scope = MediaAttachment.joins(:account).merge(Account.by_domain_and_subdomains(options[:domain]))
 | 
						|
      elsif options[:days].present?
 | 
						|
        scope = MediaAttachment.remote
 | 
						|
      else
 | 
						|
        exit(1)
 | 
						|
      end
 | 
						|
 | 
						|
      scope = scope.where('media_attachments.id > ?', Mastodon::Snowflake.id_at(options[:days].days.ago, with_random: false)) if options[:days].present?
 | 
						|
 | 
						|
      processed, aggregate = parallelize_with_progress(scope) do |media_attachment|
 | 
						|
        next if media_attachment.remote_url.blank? || (!options[:force] && media_attachment.file_file_name.present?)
 | 
						|
        next if DomainBlock.reject_media?(media_attachment.account.domain)
 | 
						|
 | 
						|
        unless options[:dry_run]
 | 
						|
          media_attachment.reset_file!
 | 
						|
          media_attachment.reset_thumbnail!
 | 
						|
          media_attachment.save
 | 
						|
        end
 | 
						|
 | 
						|
        media_attachment.file_file_size + (media_attachment.thumbnail_file_size || 0)
 | 
						|
      end
 | 
						|
 | 
						|
      say("Downloaded #{processed} media attachments (approx. #{number_to_human_size(aggregate)})#{dry_run}", :green, true)
 | 
						|
    end
 | 
						|
 | 
						|
    desc 'usage', 'Calculate disk space consumed by Mastodon'
 | 
						|
    def usage
 | 
						|
      say("Attachments:\t#{number_to_human_size(MediaAttachment.sum(Arel.sql('COALESCE(file_file_size, 0) + COALESCE(thumbnail_file_size, 0)')))} (#{number_to_human_size(MediaAttachment.where(account: Account.local).sum(Arel.sql('COALESCE(file_file_size, 0) + COALESCE(thumbnail_file_size, 0)')))} local)")
 | 
						|
      say("Custom emoji:\t#{number_to_human_size(CustomEmoji.sum(:image_file_size))} (#{number_to_human_size(CustomEmoji.local.sum(:image_file_size))} local)")
 | 
						|
      say("Preview cards:\t#{number_to_human_size(PreviewCard.sum(:image_file_size))}")
 | 
						|
      say("Avatars:\t#{number_to_human_size(Account.sum(:avatar_file_size))} (#{number_to_human_size(Account.local.sum(:avatar_file_size))} local)")
 | 
						|
      say("Headers:\t#{number_to_human_size(Account.sum(:header_file_size))} (#{number_to_human_size(Account.local.sum(:header_file_size))} local)")
 | 
						|
      say("Backups:\t#{number_to_human_size(Backup.sum(:dump_file_size))}")
 | 
						|
      say("Imports:\t#{number_to_human_size(Import.sum(:data_file_size))}")
 | 
						|
      say("Settings:\t#{number_to_human_size(SiteUpload.sum(:file_file_size))}")
 | 
						|
    end
 | 
						|
 | 
						|
    desc 'lookup URL', 'Lookup where media is displayed by passing a media URL'
 | 
						|
    def lookup(url)
 | 
						|
      path = Addressable::URI.parse(url).path
 | 
						|
 | 
						|
      path_segments = path.split('/')[2..]
 | 
						|
      path_segments.delete('cache')
 | 
						|
 | 
						|
      unless [7, 10].include?(path_segments.size)
 | 
						|
        say('Not a media URL', :red)
 | 
						|
        exit(1)
 | 
						|
      end
 | 
						|
 | 
						|
      model_name = path_segments.first.classify
 | 
						|
      record_id  = path_segments[2..-2].join.to_i
 | 
						|
 | 
						|
      unless PRELOAD_MODEL_WHITELIST.include?(model_name)
 | 
						|
        say("Cannot find corresponding model: #{model_name}", :red)
 | 
						|
        exit(1)
 | 
						|
      end
 | 
						|
 | 
						|
      record = model_name.constantize.find_by(id: record_id)
 | 
						|
      record = record.status if record.respond_to?(:status)
 | 
						|
 | 
						|
      unless record
 | 
						|
        say('Cannot find corresponding record', :red)
 | 
						|
        exit(1)
 | 
						|
      end
 | 
						|
 | 
						|
      display_url = ActivityPub::TagManager.instance.url_for(record)
 | 
						|
 | 
						|
      if display_url.blank?
 | 
						|
        say('No public URL for this type of record', :red)
 | 
						|
        exit(1)
 | 
						|
      end
 | 
						|
 | 
						|
      say(display_url, :blue)
 | 
						|
    rescue Addressable::URI::InvalidURIError
 | 
						|
      say('Invalid URL', :red)
 | 
						|
      exit(1)
 | 
						|
    end
 | 
						|
 | 
						|
    private
 | 
						|
 | 
						|
    PRELOAD_MODEL_WHITELIST = %w(
 | 
						|
      Account
 | 
						|
      Backup
 | 
						|
      CustomEmoji
 | 
						|
      Import
 | 
						|
      MediaAttachment
 | 
						|
      PreviewCard
 | 
						|
      SiteUpload
 | 
						|
    ).freeze
 | 
						|
 | 
						|
    def preload_records_from_mixed_objects(objects)
 | 
						|
      preload_map = Hash.new { |hash, key| hash[key] = [] }
 | 
						|
 | 
						|
      objects.map do |object|
 | 
						|
        segments = object.key.split('/')
 | 
						|
        segments.delete('cache')
 | 
						|
 | 
						|
        next unless [7, 10].include?(segments.size)
 | 
						|
 | 
						|
        model_name = segments.first.classify
 | 
						|
        record_id  = segments[2..-2].join.to_i
 | 
						|
 | 
						|
        next unless PRELOAD_MODEL_WHITELIST.include?(model_name)
 | 
						|
 | 
						|
        preload_map[model_name] << record_id
 | 
						|
      end
 | 
						|
 | 
						|
      preload_map.each_with_object({}) do |(model_name, record_ids), model_map|
 | 
						|
        model_map[model_name] = model_name.constantize.where(id: record_ids).index_by(&:id)
 | 
						|
      end
 | 
						|
    end
 | 
						|
  end
 | 
						|
end
 |