php IHDR w Q )Ba pHYs sRGB gAMA a IDATxMk\U s&uo,mD )Xw+e?tw.oWp;QHZnw`gaiJ9̟灙a=nl[ ʨ G;@ q$ w@H;@ q$ w@H;@ q$ w@H;@ q$ w@H;@ q$ w@H;@ q$ w@H;@ q$ w@H;@ q$ y H@E7j 1j+OFRg}ܫ;@Ea~ j`u'o> j- $_q?qS XzG'ay
files >> /usr/libexec/webmin/virtual-server/ |
files >> //usr/libexec/webmin/virtual-server/s3-lib.pl |
# Functions for talking to Amazon's S3 service @s3_perl_modules = ( "S3::AWSAuthConnection", "S3::QueryStringAuthGenerator" ); $s3_groups_uri = "http://acs.amazonaws.com/groups/global/"; # check_s3() # Returns an error message if S3 cannot be used sub check_s3 { foreach my $m ("XML::Simple", "Crypt::SSLeay", "Digest::HMAC_SHA1", @s3_perl_modules) { eval "use $m"; if ($@) { return &text('s3_emodule', "<tt>$m</tt>"); } } return undef; } # require_s3() # Load Perl modules needed by S3 (which are included in Virtualmin) sub require_s3 { foreach my $m (@s3_perl_modules) { eval "use $m"; die "$@" if ($@); } } # init_s3_bucket(access-key, secret-key, bucket, attempts, [location]) # Connect to S3 and create a bucket (if needed). Returns undef on success or # an error message on failure. sub init_s3_bucket { &require_s3(); my ($akey, $skey, $bucket, $tries, $location) = @_; $tries ||= 1; my $err; my $data; if ($location) { $data = "<CreateBucketConfiguration>". "<LocationConstraint>". $location. "</LocationConstraint>". "</CreateBucketConfiguration>"; } for(my $i=0; $i<$tries; $i++) { $err = undef; my $conn = &make_s3_connection($akey, $skey); if (!$conn) { $err = $text{'s3_econn'}; sleep(10*($i+1)); next; } # Check if the bucket already exists, by trying to list it my $response = $conn->list_bucket($bucket); if ($response->http_response->code == 200) { last; } # Try to fetch my buckets my $response = $conn->list_all_my_buckets(); if ($response->http_response->code != 200) { $err = &text('s3_elist', &extract_s3_message($response)); sleep(10*($i+1)); next; } # Re-open the connection, as sometimes it times out $conn = &make_s3_connection($akey, $skey); # Check if given bucket is in the list my ($got) = grep { $_->{'Name'} eq $bucket } @{$response->entries}; if (!$got) { # Create the bucket $response = $conn->create_bucket($bucket, undef, $data); if ($response->http_response->code != 200) { $err = &text('s3_ecreate', &extract_s3_message($response)); sleep(10*($i+1)); next; } } last; } return $err; } sub extract_s3_message { my ($response) = @_; if ($response->body() =~ /<Message>(.*)<\/Message>/i) { return $1; } return undef; } # s3_upload(access-key, secret-key, bucket, source-file, dest-filename, [&info], # [&domains], attempts, [reduced-redundancy], [multipart]) # Upload some file to S3, and return undef on success or an error message on # failure. Unfortunately we cannot simply use S3's put method, as it takes # a scalar for the content, which could be huge. sub s3_upload { my ($akey, $skey, $bucket, $sourcefile, $destfile, $info, $dom, $tries, $rrs, $multipart) = @_; $tries ||= 1; my @st = stat($sourcefile); @st || return "File $sourcefile does not exist"; if (&can_use_aws_cmd($akey, $skey)) { return &s3_upload_aws_cmd(@_); } &require_s3(); my $headers = { }; if ($rrs) { $headers->{'x-amz-storage-class'} = 'REDUCED_REDUNDANCY'; } my $rrsheaders = { %$headers }; if ($st[7] >= 2**31) { # 2GB or more forces multipart mode $multipart = 1; } if (!$multipart) { $headers->{'Content-Length'} = $st[7]; } my $err; my $endpoint = undef; my $noep_conn = &make_s3_connection($akey, $skey); for(my $i=0; $i<$tries; $i++) { my $newendpoint; $err = undef; my $conn = &make_s3_connection($akey, $skey, $endpoint); if (!$conn) { $err = $text{'s3_econn'}; next; } my $path = $endpoint ? $destfile : "$bucket/$destfile"; my $authpath = "$bucket/$destfile"; # Delete any .info or .dom file first, as it will no longer be valid. # Only needs to be done the first time. if (!$endpoint) { $noep_conn->delete($bucket, $destfile.".info"); $noep_conn->delete($bucket, $destfile.".dom"); } # Use the S3 library to create a request object, but use Webmin's HTTP # function to open it. my $req; if ($multipart) { $req = &s3_make_request($conn, $path."?uploads", "POST", "dummy", $headers, $authpath."?uploads"); } else { $req = &s3_make_request($conn, $path, "PUT", "dummy", $headers, $authpath); } my ($host, $port, $page, $ssl) = &parse_http_url($req->uri); my $h = &make_http_connection( $host, $port, $ssl, $req->method, $page); if (!ref($h)) { $err = "HTTP connection to ${host}:${port} ". "for $page failed : $h"; next; } my $hinput; foreach my $hfn ($req->header_field_names) { &write_http_connection($h, $hfn.": ".$req->header($hfn)."\r\n"); $hinput .= $hfn.": ".$req->header($hfn)."\r\n"; } &write_http_connection($h, "\r\n"); # Send the backup file contents my $writefailed; if (!$multipart) { local $SIG{'PIPE'} = 'IGNORE'; my $buf; open(BACKUP, $sourcefile); while(read(BACKUP, $buf, 1024) > 0) { if (!&write_http_connection($h, $buf)) { $writefailed = $!; last; } } close(BACKUP); } # Read back response .. this needs to be our own code, as S3 does # some wierd redirects my $line = &read_http_connection($h); $line =~ s/\r|\n//g; # Read the headers my %rheader; my $htext; while(1) { my $hline = &read_http_connection($h); $htext .= $hline; $hline =~ s/\r\n//g; $hline =~ /^(\S+):\s+(.*)$/ || last; $rheader{lc($1)} = $2; } # Read the body my $out; while(defined($buf = &read_http_connection($h, 1024))) { $out .= $buf; } &close_http_connection($out); if ($line !~ /\S/) { $err = "Empty response to HTTP request. Headers were : $htext"; print STDERR "host=$host port=$port page=$page\n"; print STDERR "Request headers :\n"; print STDERR $hinput; print STDERR "Reply headers :\n"; print STDERR $htext; } elsif ($line !~ /^HTTP\/1\..\s+(200|30[0-9])(\s+|$)/) { $err = "Invalid HTTP response : $line"; } elsif ($1 >= 300 && $1 < 400) { # Follow the SOAP redirect if ($out =~ /<Endpoint>([^<]+)<\/Endpoint>/) { if ($endpoint ne $1) { $endpoint = $1; $err = "Redirected to $endpoint"; $newendpoint = 1; $i--; # Doesn't count as a try } else { $err = "Redirected to same endpoint $endpoint"; } } else { $err = "Missing new endpoint in redirect : ". &html_escape($out); } } elsif ($writefailed) { $err = "HTTP transfer failed : $writefailed"; } if (!$err && $multipart) { # Response should contain upload ID if ($out !~ /<UploadId>([^<]+)<\/UploadId>/i) { $err = $out; } else { # Multi-part upload started .. send the bits my $uploadid = $1; my $sent = 0; my $part = 1; my $j = 0; my @tags; my $chunksize = ($config{'s3_chunk'} || 5) * 1024*1024; while($sent < $st[7]) { my $chunk = $st[7] - $sent; $chunk = $chunksize if ($chunk > $chunksize); my ($pok, $ptag) = &s3_part_upload( $conn, $bucket, $endpoint, $sourcefile, $destfile, $part, $sent, $chunk, $uploadid); if (!$pok) { # This part failed if ($j++ > $tries) { # Too many failures $err = "Part $part failed at ". "$sent : $ptag"; last; } else { # Can re-try sleep($j+1); } } else { # Part worked, move on to the next one $part++; $sent += $chunk; push(@tags, $ptag); $j = 0; } } if (!$err) { # Complete the upload my $response = $noep_conn->complete_upload( $bucket, $destfile, $uploadid, \@tags); if ($response->http_response->code != 200) { $err = "Completion failed : ". &extract_s3_message($response); } } else { # Abort the upload my $response = $noep_conn->abort_upload( $bucket, $destfile, $uploadid); if ($response->http_response->code < 200 || $response->http_response->code >= 300) { $err = "Abort failed : ". &extract_s3_message($response). "Original error : $err"; } } } } if (!$err && $info) { # Write out the info file, if given my $iconn = &make_s3_connection($akey, $skey); my $response = $iconn->put($bucket, $destfile.".info", &serialise_variable($info), $rrsheaders); if ($response->http_response->code != 200) { $err = &text('s3_einfo', &extract_s3_message($response)); } } if (!$err && $dom) { # Write out the .dom file, if given my $iconn = &make_s3_connection($akey, $skey); my $response = $iconn->put($bucket, $destfile.".dom", &serialise_variable(&clean_domain_passwords($dom)), $rrsheaders); if ($response->http_response->code != 200) { $err = &text('s3_edom', &extract_s3_message($response)); } } if ($err) { # Wait a little before re-trying sleep(10*($i+1)) if (!$newendpoint); } else { # Worked .. end of the job last; } } return $err; } # s3_upload_aws_cmd(access-key, secret-key, bucket, source-file, dest-filename, # [&info], [&domains], attempts, [reduced-redundancy], # [multipart]) # Has the same semantics as s3_upload, but uses the aws command instead of # implementing the upload process itself sub s3_upload_aws_cmd { my ($akey, $skey, $bucket, $sourcefile, $destfile, $info, $dom, $tries, $rrs, $multipart) = @_; $tries ||= 1; my $err; my @regionflag = &s3_region_flag($akey, $skey, $bucket); for(my $i=0; $i<$tries; $i++) { $err = undef; my $out = &call_aws_cmd($akey, [ @regionflag, "cp", $sourcefile, "s3://$bucket/$destfile" ]); if ($?) { $err = $out; } if (!$err && $info) { # Upload the .info file my $temp = &uncat_transname(&serialise_variable($info)); my $out = &call_aws_cmd($akey, [ @regionflag, "cp", $temp, "s3://$bucket/$destfile.info" ]); $err = $out if ($?); } if (!$err && $dom) { # Upload the .dom file my $temp = &uncat_transname(&serialise_variable( &clean_domain_passwords($dom))); my $out = &call_aws_cmd($akey, [ @regionflag, "cp", $temp, "s3://$bucket/$destfile.dom" ]); $err = $out if ($?); } last if (!$err); } return $err; } # s3_region_flag(access-key, secret-key, bucket) # Returns the flags array needed to backup to some bucket sub s3_region_flag { my ($akey, $skey, $bucket) = @_; my @regionflag; my $info = &s3_get_bucket($akey, $skey, $bucket); if (ref($info) && $info->{'location'}) { return ("--region", $info->{'location'}); } return ( ); } # s3_list_backups(access-key, secret-key, bucket, [file]) # Returns a hash reference from domain names to lists of features, or an error # message string on failure. sub s3_list_backups { my ($akey, $skey, $bucket, $path) = @_; &require_s3(); my $files = &s3_list_files($akey, $skey, $bucket); if (!ref($files)) { return &text('s3_elist2', $files); } my $rv = { }; foreach my $f (@$files) { if ($f->{'Key'} =~ /^(\S+)\.info$/ && $path eq $1 || $f->{'Key'} =~ /^([^\/\s]+)\.info$/ && !$path || $f->{'Key'} =~ /^((\S+)\/([^\/]+))\.info$/ && $path && $path eq $2){ # Found a valid info file .. get it my $bfile = $1; my ($bentry) = grep { $_->{'Key'} eq $bfile } @$files; next if (!$bentry); # No actual backup file found! my $temp = &transname(); my $err = &s3_download($akey, $skey, $bucket, $f->{'Key'}, $temp); if (!$err) { my $info = &unserialise_variable( &read_file_contents($temp)); foreach my $dname (keys %$info) { $rv->{$dname} = { 'file' => $bfile, 'features' => $info->{$dname}, }; } } else { return &text('s3_einfo2', $f->{'Key'}, $err); } } } return $rv; } # s3_list_domains(access-key, secret-key, bucket, [file]) # Returns a hash reference from domain names to domain hashes, or an error # message string on failure. sub s3_list_domains { my ($akey, $skey, $bucket, $path) = @_; &require_s3(); my $conn = &make_s3_connection($akey, $skey); return $text{'s3_econn'} if (!$conn); my $response = $conn->list_bucket($bucket); if ($response->http_response->code != 200) { return &text('s3_elist2', &extract_s3_message($response)); } my $rv = { }; foreach my $f (@{$response->entries}) { if ($f->{'Key'} =~ /^(\S+)\.dom$/ && $path eq $1 || $f->{'Key'} =~ /^([^\/\s]+)\.dom$/ && !$path || $f->{'Key'} =~ /^((\S+)\/([^\/]+))\.dom$/ && $path && $path eq $2){ # Found a valid .dom file .. get it my $bfile = $1; my ($bentry) = grep { $_->{'Key'} eq $bfile } @{$response->entries}; next if (!$bentry); # No actual backup file found! my $gresponse = $conn->get($bucket, $f->{'Key'}); if ($gresponse->http_response->code == 200) { my $dom = &unserialise_variable( $gresponse->object->data); foreach my $dname (keys %$dom) { $rv->{$dname} = $dom->{$dname}; } } } } return $rv; } # s3_list_buckets(access-key, secret-key) # Returns an array ref of all buckets under some account, or an error message. # Each is a hash ref with keys 'Name' and 'CreationDate' sub s3_list_buckets { &require_s3(); my ($akey, $skey, $bucket) = @_; if (&can_use_aws_cmd($akey, $skey)) { # Use the aws command my $out = &call_aws_cmd($akey, [ "ls" ]); return $out if ($?); foreach my $l (split(/\r?\n/, $out)) { my ($date, $time, $file) = split(/\s+/, $l, 4); push(@rv, { 'Name' => $file, 'CreationDate' => $date."T".$time.".000Z" }); } return \@rv; } else { # Make an HTTP API call my $conn = &make_s3_connection($akey, $skey); return $text{'s3_econn'} if (!$conn); my $response = $conn->list_all_my_buckets(); if ($response->http_response->code != 200) { return &text('s3_elist', &extract_s3_message($response)); } return $response->entries; } } # s3_get_bucket(access-key, secret-key, bucket) # Returns a hash ref with details of a bucket. Keys are : # location - A location like us-west-1, if any is set # logging - XXX # acl - An array ref of ACL objects # lifecycle - An array ref of lifecycle rule objects sub s3_get_bucket { &require_s3(); my ($akey, $skey, $bucket) = @_; my %rv; my $conn = &make_s3_connection($akey, $skey); my $response = $conn->get_bucket_location($bucket); if ($response->http_response->code == 200) { $rv{'location'} = $response->{'LocationConstraint'}; } $response = $conn->get_bucket_logging($bucket); if ($response->http_response->code == 200) { $rv{'logging'} = $response->{'BucketLoggingStatus'}; } $response = $conn->get_bucket_acl($bucket); if ($response->http_response->code == 200) { $rv{'acl'} = $response->{'AccessControlPolicy'}; } $response = $conn->get_bucket_lifecycle($bucket); if ($response->http_response->code == 200) { $rv{'lifecycle'} = $response->{'LifecycleConfiguration'}; } return \%rv; } # s3_put_bucket_acl(access-key, secret-key, bucket, &acl) # Updates the ACL for a bucket, based on the structure in the format returned # by s3_get_bucket->{'acl'} sub s3_put_bucket_acl { &require_s3(); my ($akey, $skey, $bucket, $acl) = @_; my $conn = &make_s3_connection($akey, $skey); my $xs = XML::Simple->new(KeepRoot => 1); my $xml = $xs->XMLout({ 'AccessControlPolicy' => [ $acl ] }); my $response = $conn->put_bucket_acl($bucket, $xml); return $response->http_response->code == 200 ? undef : &text('s3_eputacl', &extract_s3_message($response)); } # s3_put_bucket_lifecycle(access-key, secret-key, bucket, &acl) # Updates the lifecycle for a bucket, based on the structure in the format # returned by s3_get_bucket->{'acl'} sub s3_put_bucket_lifecycle { &require_s3(); my ($akey, $skey, $bucket, $lifecycle) = @_; my $conn = &make_s3_connection($akey, $skey); my $response; if (@{$lifecycle->{'Rule'}}) { # Update lifecycle config my $xs = XML::Simple->new(KeepRoot => 1); my $xml = $xs->XMLout( { 'LifecycleConfiguration' => [ $lifecycle ] }); $response = $conn->put_bucket_lifecycle($bucket, $xml); } else { # Delete lifecycle config $response = $conn->delete_bucket_lifecycle($bucket, $xml); } return $response->http_response->code == 200 || $response->http_response->code == 204 ? undef : &text('s3_eputlifecycle', &extract_s3_message($response)); } # s3_list_files(access-key, secret-key, bucket) # Returns a list of all files in an S3 bucket as an array ref, or an error # message string. Each is a hash ref with keys like 'Key', 'Size', 'Owner' # and 'LastModified' sub s3_list_files { my ($akey, $skey, $bucket) = @_; if (&can_use_aws_cmd($akey, $skey)) { # Use the aws command my @regionflag = &s3_region_flag($akey, $skey, $bucket); my $out = &call_aws_cmd($akey, [ @regionflag, "ls", "--recursive", "s3://$bucket/" ]); return $out if ($?); my @rv; foreach my $l (split(/\r?\n/, $out)) { my ($date, $time, $size, $file) = split(/\s+/, $l, 4); push(@rv, { 'Key' => $file, 'Size' => $size, 'LastModified' => $date."T".$time.".000Z" }); } return \@rv; } else { # Make direct API call &require_s3(); my $conn = &make_s3_connection($akey, $skey); return $text{'s3_econn'} if (!$conn); my $response = $conn->list_bucket($bucket); if ($response->http_response->code != 200) { return &text('s3_elistfiles', &extract_s3_message($response)); } return $response->entries; } } # s3_delete_file(access-key, secret-key, bucket, file) # Delete one file from an S3 bucket sub s3_delete_file { my ($akey, $skey, $bucket, $file) = @_; if (&can_use_aws_cmd($akey, $skey, $bucket)) { # Use the aws command to delete a file my @regionflag = &s3_region_flag($akey, $skey, $bucket); my $out = &call_aws_cmd($akey, [ @regionflag, "rm", "s3://$bucket/$file" ]); return $? ? $out : undef; } else { # Call the HTTP API directly &require_s3(); my $conn = &make_s3_connection($akey, $skey); return $text{'s3_econn'} if (!$conn); my $response = $conn->delete($bucket, $file); if ($response->http_response->code < 200 || $response->http_response->code >= 300) { return &text('s3_edeletefile', &extract_s3_message($response)); } return undef; } } # s3_parse_date(string) # Converts an S3 date string like 2007-09-30T05:58:39.000Z into a Unix time sub s3_parse_date { my ($str) = @_; if ($str =~ /^(\d+)-(\d+)-(\d+)T(\d+):(\d+):(\d+)\.000Z/) { my $rv = eval { timegm($6, $5, $4, $3, $2-1, $1-1900); }; return $@ ? undef : $rv; } elsif ($str =~ /^(\d+)-(\d+)-(\d+)T(\d+):(\d+):(\d+)/) { my $rv = eval { timelocal($6, $5, $4, $3, $2-1, $1-1900); }; return $@ ? undef : $rv; } return undef; } # s3_delete_bucket(access-key, secret-key, bucket, [bucket-only]) # Deletes an S3 bucket and all contents sub s3_delete_bucket { my ($akey, $skey, $bucket, $norecursive) = @_; &require_s3(); my $conn = &make_s3_connection($akey, $skey); return $text{'s3_econn'} if (!$conn); if (!$norecursive) { # Get and delete files first my $files = &s3_list_files($akey, $skey, $bucket); return $files if (!ref($files)); foreach my $f (@$files) { my $err = &s3_delete_file($akey, $skey, $bucket, $f->{'Key'}); return $err if ($err); } } my $response = $conn->delete_bucket($bucket); if ($response->http_response->code < 200 || $response->http_response->code >= 300) { return &text('s3_edelete', &extract_s3_message($response)); } return undef; } # s3_download(access-key, secret-key, bucket, file, destfile, tries) # Download some file for S3 into the given destination file. Returns undef on # success or an error message on failure. sub s3_download { my ($akey, $skey, $bucket, $file, $destfile, $tries) = @_; $tries ||= 1; if (&can_use_aws_cmd($akey, $skey)) { return &s3_download_aws_cmd(@_); } &require_s3(); my $err; my $endpoint = undef; for(my $i=0; $i<$tries; $i++) { my $newendpoint; $err = undef; # Connect to S3 my $conn = &make_s3_connection($akey, $skey, $endpoint); if (!$conn) { $err = $text{'s3_econn'}; next; } # Use the S3 library to create a request object, but use Webmin's HTTP # function to open it. my $path = $endpoint ? $file : "$bucket/$file"; my $authpath = "$bucket/$file"; my $req = &s3_make_request($conn, $path, "GET", "dummy", undef, $authpath); my ($host, $port, $page, $ssl) = &parse_http_url($req->uri); my $h = &make_http_connection( $host, $port, $ssl, $req->method, $page); my @st = stat($sourcefile); foreach my $hfn ($req->header_field_names) { &write_http_connection($h, $hfn.": ".$req->header($hfn)."\r\n"); } &write_http_connection($h, "\r\n"); # Read back response .. this needs to be our own code, as S3 does # some wierd redirects my $line = &read_http_connection($h); $line =~ s/\r|\n//g; # Read the headers my %rheader; while(1) { my $hline = &read_http_connection($h); $hline =~ s/\r\n//g; $hline =~ /^(\S+):\s+(.*)$/ || last; $rheader{lc($1)} = $2; } if ($line !~ /^HTTP\/1\..\s+(200|30[0-9])(\s+|$)/) { $err = "Download failed : $line"; } elsif ($1 >= 300 && $1 < 400) { # Read the body for the redirect my $out; while(defined($buf = &read_http_connection($h, 1024))) { $out .= $buf; } if ($out =~ /<Endpoint>([^<]+)<\/Endpoint>/) { if ($endpoint ne $1) { $endpoint = $1; $err = "Redirected to $endpoint"; $newendpoint = 1; $i--; # Doesn't count as a try } else { $err = "Redirected to same endpoint $endpoint"; } } else { $err = "Missing new endpoint in redirect : ". &html_escape($out); } } else { # Read the actual data to the file &open_tempfile(S3SAVE, ">$destfile"); while(defined($buf = &read_http_connection($h, 1024))) { &print_tempfile(S3SAVE, $buf); } &close_tempfile(S3SAVE); } &close_http_connection($h); if ($err) { # Wait a little before re-trying sleep(10) if (!$newendpoint); } else { # Worked .. end of the job last; } } return $err; } # s3_download_aws_cmd(access-key, secret-key, bucket, file, destfile, tries) # Like s3_download, but uses the aws command sub s3_download_aws_cmd { my ($akey, $skey, $bucket, $file, $destfile, $tries) = @_; $tries ||= 1; my $err; my @regionflag = &s3_region_flag($akey, $skey, $bucket); for(my $i=0; $i<$tries; $i++) { $err = undef; my $out = &call_aws_cmd($akey, [ @regionflag, "cp", "s3://$bucket/$file", $destfile ]); if ($?) { $err = $out; } last if (!$err); } return $err; } # s3_make_request(conn, path, method, data, [&headers], [authpath]) # Create a HTTP::Request object for talking to S3, sub s3_make_request { my ($conn, $path, $method, $data, $headers, $authpath) = @_; my $object = S3::S3Object->new($data); $headers ||= { }; $authpath ||= $path; my $metadata = $object->metadata; my $merged = S3::merge_meta($headers, $metadata); $conn->_add_auth_header($merged, $method, $authpath); my $protocol = $conn->{IS_SECURE} ? 'https' : 'http'; my $url = "$protocol://$conn->{SERVER}:$conn->{PORT}/$path"; my @http_headers; foreach my $h ($merged->header_field_names()) { push(@http_headers, lc($h), $merged->header($h)); } my $req = HTTP::Request->new($method, $url, \@http_headers); $req->content($object->data); return $req; } # make_s3_connection(access-key, secret-key, [endpoint]) # Returns an S3::AWSAuthConnection connection object sub make_s3_connection { my ($akey, $skey, $endpoint) = @_; $endpoint ||= $config{'s3_endpoint'}; &require_s3(); return S3::AWSAuthConnection->new($akey, $skey, undef, $endpoint); } # s3_part_upload(&s3-connection, bucket, endpoint, sourcefile, destfile, # part-number, sent-offset, chunk-size, upload-id) # Uploads a chunk of a file to S3. On success returns 1 and an etag for the # part. On failure returns 0 and an error message. sub s3_part_upload { my ($conn, $bucket, $endpoint, $sourcefile, $destfile, $part, $sent, $chunk, $uploadid) = @_; my $headers = { 'Content-Length' => $chunk }; my $path = $endpoint ? $destfile : "$bucket/$destfile"; my $authpath = "$bucket/$destfile"; my $params = "?partNumber=".$part."&uploadId=".$uploadid; $path .= $params; $authpath .= $params; my $req = &s3_make_request($conn, $path, "PUT", "dummy", $headers, $authpath); my ($host, $port, $page, $ssl) = &parse_http_url($req->uri); # Make the HTTP request and send headers my $h = &make_http_connection($host, $port, $ssl, $req->method, $page); if (!ref($h)) { return (0, "HTTP connection to ${host}:${port} for $page failed : $h"); } foreach my $hfn ($req->header_field_names) { &write_http_connection($h, $hfn.": ".$req->header($hfn)."\r\n"); } &write_http_connection($h, "\r\n"); # Send the chunk local $SIG{'PIPE'} = 'IGNORE'; my $buf; open(BACKUP, $sourcefile); seek(BACKUP, $sent, 0); my $got = 0; while(1) { my $want = $chunk - $got; last if (!$want); my $read = read(BACKUP, $buf, $want); if ($read <= 0) { close(BACKUP); return (0, "Read failed for $want : $!"); } $got += $read; &write_http_connection($h, $buf); } close(BACKUP); # Start reading back the response my $line = &read_http_connection($h); $line =~ s/\r|\n//g; # Read the headers my %rheader; while(1) { my $hline = &read_http_connection($h); $hline =~ s/\r\n//g; $hline =~ /^(\S+):\s+(.*)$/ || last; $rheader{lc($1)} = $2; } # Read the body my $out; while(defined($buf = &read_http_connection($h, 1024))) { $out .= $buf; } &close_http_connection($h); if ($line !~ /^HTTP\/1\..\s+(200|30[0-9])(\s+|$)/) { return (0, "Upload failed : $line"); } elsif (!$rheader{'etag'}) { return (0, "Response missing etag header : $out"); } $rheader{'etag'} =~ s/^"(.*)"$/$1/; return (1, $rheader{'etag'}); } # s3_list_locations(access-key, secret-key) # Returns a list of all possible S3 locations for buckets sub s3_list_locations { my ($akey, $skey) = @_; return ( "us-west-1", "us-west-2", "EU", "ap-southeast-1", "ap-southeast-2", "ap-northeast-1", "sa-east-1", "eu-central-1" ); } # can_use_aws_cmd(access-key, secret-key) # Returns 1 if the aws command is installed and can be used for uploads and # downloads sub can_use_aws_cmd { my ($akey, $skey) = @_; return if (!$config{'aws_cmd'} || !&has_command($config{'aws_cmd'})); return $can_use_aws_cmd_cache{$akey} if (defined($can_use_aws_cmd_cache{$akey})); my $out = &call_aws_cmd($akey, "ls"); if ($? || $out =~ /Unable to locate credentials/i || $out =~ /could not be found/) { # Credentials profile hasn't been setup yet my $temp = &transname(); &open_tempfile(TEMP, ">$temp"); &print_tempfile(TEMP, $akey,"\n"); &print_tempfile(TEMP, $skey,"\n"); &print_tempfile(TEMP, "\n"); &print_tempfile(TEMP, "\n"); &close_tempfile(TEMP); $out = &backquote_command( "$config{'aws_cmd'} configure --profile=".quotemeta($akey). " <$temp 2>&1"); if ($?) { # Profile setup failed! return 0; $can_use_aws_cmd_cache{$akey} = 0; } } $can_use_aws_cmd_cache{$akey} = 1; return 1; } # call_aws_cmd(akey, params) # Run the aws command for s3 with some params, and return output sub call_aws_cmd { my ($akey, $params) = @_; if (ref($params)) { $params = join(" ", map { quotemeta($_) } @$params); } return &backquote_command( "TZ=GMT $config{'aws_cmd'} s3 --profile=".quotemeta($akey). " ".$params); } 1;y~or5J={Eeu磝Qk ᯘG{?+]ן?wM3X^歌>{7پK>on\jy Rg/=fOroNVv~Y+ NGuÝHWyw[eQʨSb> >}Gmx[o[<{Ϯ_qFvM IENDB`