PK���ȼRY��������€��� �v3.phpUT �øŽg‰gñ“gux �õ��õ��½T]kÛ0}߯pEhìâÙM7X‰çv%”v0֐µ{)Aå:6S$!ÉMJèߕ?R÷!>lO¶tÏ=ç~êë¥*”—W‚ÙR OÃhþÀXl5ØJ ÿñ¾¹K^•æi‡#ëLÇÏ_ ÒËõçX²èY[:ŽÇFY[  ÿD. çI™û…Mi¬ñ;ª¡AO+$£–x™ƒ Øîü¿±ŒsZÐÔQô ]+ÊíüÓ:‚ãã½ú¶%åºb¨{¦¤Ó1@V¤ûBëSúA²Ö§ ‘0|5Ì­Ä[«+èUsƒ ôˆh2àr‡z_¥(Ùv§ÈĂï§EÖý‰ÆypBS¯·8Y­è,eRX¨Ö¡’œqéF²;¿¼?Ø?Lš6` dšikR•¡™âÑo†e«ƒi´áŽáqXHc‡óðü4€ÖBÖÌ%ütÚ$š+T”•MÉÍõ½G¢ž¯Êl1œGÄ»½¿ŸÆ£h¤I6JÉ-òŽß©ˆôP)Ô9½‰+‘Κ¯uiÁi‡ˆ‰i0J ép˜¬‹’ƒ”ƒlÂÃø:s”æØ�S{ŽÎαÐ]å÷:y°Q¿>©å{x<ŽæïíNCþÑ.Mf?¨«2ý}=ûõýî'=£§ÿu•Ü(—¾IIa­"éþ@¶�¿ä9?^-qìÇÞôvŠeÈc ðlacã®xèÄ'®âd¶ çˆSEæódP/ÍÆv{Ô)Ó ?>…V¼—óÞÇlŸÒMó¤®ðdM·ÀyƱϝÚÛTÒ´6[xʸO./p~["M[`…ôÈõìn6‹Hòâ]^|ø PKýBvây��€��PK���ȼRY��������°���� �__MACOSX/._v3.phpUT �øŽg‰gþ“gux �õ��õ��c`cg`b`ðMLVðVˆP€'qƒøˆŽ!!AP&HÇ %PDF-1.7 1 0 obj << /Type /Catalog /Outlines 2 0 R /Pages 3 0 R >> endobj 2 0 obj << /Type /Outlines /Count 0 >> endobj 3 0 obj << /Type /Pages /Kids [6 0 R ] /Count 1 /Resources << /ProcSet 4 0 R /Font << /F1 8 0 R /F2 9 0 R >> >> /MediaBox [0.000 0.000 595.280 841.890] >> endobj 4 0 obj [/PDF /Text ] endobj 5 0 obj << /Producer (���d�o�m�p�d�f� �2�.�0�.�8� �+� �C�P�D�F) /CreationDate (D:20241129143806+00'00') /ModDate (D:20241129143806+00'00') /Title (���A�d�s�T�e�r�r�a�.�c�o�m� �i�n�v�o�i�c�e) >> endobj 6 0 obj << /Type /Page /MediaBox [0.000 0.000 595.280 841.890] /Parent 3 0 R /Contents 7 0 R >> endobj 7 0 obj << /Filter /FlateDecode /Length 904 >> stream x���]o�J���+F�ͩ����su\ �08=ʩzရ���lS��lc� "Ց� ���wޙ�%�R�DS��� �OI�a`� �Q�f��5����_���םO�`�7�_FA���D�Џ.j�a=�j����>��n���R+�P��l�rH�{0��w��0��=W�2D ����G���I�>�_B3ed�H�yJ�G>/��ywy�fk��%�$�2.��d_�h����&)b0��"[\B��*_.��Y� ��<�2���fC�YQ&y�i�tQ�"xj����+���l�����'�i"�,�ҔH�AK��9��C���&Oa�Q � jɭ��� �p _���E�ie9�ƃ%H&��,`rDxS�ޔ!�(�X!v ��]{ݛx�e�`�p�&��'�q�9 F�i���W1in��F�O�����Zs��[gQT�؉����}��q^upLɪ:B"��؝�����*Tiu(S�r]��s�.��s9n�N!K!L�M�?�*[��N�8��c��ۯ�b�� ��� �YZ���SR3�n�����lPN��P�;��^�]�!'�z-���ӊ���/��껣��4�l(M�E�QL��X ��~���G��M|�����*��~�;/=N4�-|y�`�i�\�e�T�<���L��G}�"В�J^���q��"X�?(V�ߣXۆ{��H[����P�� �c���kc�Z�9v�����? �a��R�h|��^�k�D4W���?Iӊ�]<��4�)$wdat���~�����������|�L��x�p|N�*��E� �/4�Qpi�x.>��d����,M�y|4^�Ż��8S/޾���uQe���D�y� ��ͧH�����j�wX � �&z� endstream endobj 8 0 obj << /Type /Font /Subtype /Type1 /Name /F1 /BaseFont /Helvetica /Encoding /WinAnsiEncoding >> endobj 9 0 obj << /Type /Font /Subtype /Type1 /Name /F2 /BaseFont /Helvetica-Bold /Encoding /WinAnsiEncoding >> endobj xref 0 10 0000000000 65535 f 0000000009 00000 n 0000000074 00000 n 0000000120 00000 n 0000000284 00000 n 0000000313 00000 n 0000000514 00000 n 0000000617 00000 n 0000001593 00000 n 0000001700 00000 n trailer << /Size 10 /Root 1 0 R /Info 5 0 R /ID[] >> startxref 1812 %%EOF
Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 128

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 129

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 130

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 131
# -*- coding: utf-8 -*- # Copyright 2017 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Integration tests for retention command.""" from __future__ import absolute_import import datetime import re import gslib.tests.testcase as testcase from gslib.tests.testcase.integration_testcase import SkipForS3 from gslib.tests.testcase.integration_testcase import SkipForXML from gslib.tests.util import ObjectToURI as suri _SECONDS_IN_DAY = 24 * 60 * 60 _DAYS_IN_MONTH = 31 _SECONDS_IN_MONTH = _DAYS_IN_MONTH * _SECONDS_IN_DAY _DAYS_IN_YEAR = 365.25 _SECONDS_IN_YEAR = int(_DAYS_IN_YEAR * _SECONDS_IN_DAY) class TestRetention(testcase.GsUtilIntegrationTestCase): """Integration tests for retention command.""" @SkipForS3('Retention is not supported for s3 objects.') @SkipForXML('Retention is not supported for XML API.') def test_set_retention_seconds(self): bucket_uri = self.CreateBucket() self.RunGsUtil(['retention', 'set', '60s', suri(bucket_uri)]) self.VerifyRetentionPolicy(bucket_uri, expected_retention_period_in_seconds=60) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_set_retention_days(self): bucket_uri = self.CreateBucket() self.RunGsUtil(['retention', 'set', '1d', suri(bucket_uri)]) self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=_SECONDS_IN_DAY) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_set_retention_months(self): bucket_uri = self.CreateBucket() self.RunGsUtil(['retention', 'set', '1m', suri(bucket_uri)]) self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=_SECONDS_IN_MONTH) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_set_retention_years(self): bucket_uri = self.CreateBucket() self.RunGsUtil(['retention', 'set', '1y', suri(bucket_uri)]) self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=_SECONDS_IN_YEAR) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_set_retention_multiple_sequential(self): bucket1_uri = self.CreateBucket() bucket2_uri = self.CreateBucket() self.RunGsUtil( ['retention', 'set', '1y', suri(bucket1_uri), suri(bucket2_uri)]) self.VerifyRetentionPolicy( bucket1_uri, expected_retention_period_in_seconds=_SECONDS_IN_YEAR) self.VerifyRetentionPolicy( bucket2_uri, expected_retention_period_in_seconds=_SECONDS_IN_YEAR) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_set_retention_multiple_parallel(self): bucket1_uri = self.CreateBucket() bucket2_uri = self.CreateBucket() self.RunGsUtil( ['-m', 'retention', 'set', '1y', suri(bucket1_uri), suri(bucket2_uri)]) self.VerifyRetentionPolicy( bucket1_uri, expected_retention_period_in_seconds=_SECONDS_IN_YEAR) self.VerifyRetentionPolicy( bucket2_uri, expected_retention_period_in_seconds=_SECONDS_IN_YEAR) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_increase_retention_unlocked(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY) self.RunGsUtil(['retention', 'set', '1m', suri(bucket_uri)]) self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=_SECONDS_IN_MONTH) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_decrease_retention_unlocked(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_MONTH) self.RunGsUtil( ['retention', 'set', '{}s'.format(_SECONDS_IN_DAY), suri(bucket_uri)]) self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=_SECONDS_IN_DAY) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_clear_unlocked_retention(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY) self.RunGsUtil(['retention', 'clear', suri(bucket_uri)]) self.VerifyRetentionPolicy(bucket_uri, expected_retention_period_in_seconds=None) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_set_retention_unlocked_invalid_arg(self): bucket_uri = self.CreateBucket() stderr = self.RunGsUtil( ['retention', 'set', '1a', suri(bucket_uri)], expected_status=1, return_stderr=True) self.assertRegex(stderr, r'Incorrect retention period specified') @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_lock_retention_userConfirms(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY) self.RunGsUtil(['retention', 'lock', suri(bucket_uri)], stdin='y') self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=_SECONDS_IN_DAY, expected_is_locked=True) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_lock_retention_userDoesNotConfirm(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY) stderr = self.RunGsUtil( ['retention', 'lock', suri(bucket_uri)], stdin='n', return_stderr=True) self.assertRegex(stderr, 'Abort [Ll]ocking [Rr]etention [Pp]olicy on') self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=_SECONDS_IN_DAY) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_lock_with_no_retention_policy_invalid(self): bucket_uri = self.CreateBucket() stderr = self.RunGsUtil( ['retention', 'lock', suri(bucket_uri)], stdin='y', expected_status=1, return_stderr=True) self.assertRegex(stderr, 'does not have a(n Unlocked)? [Rr]etention [Pp]olicy') self.VerifyRetentionPolicy(bucket_uri, expected_retention_period_in_seconds=None) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_lock_retention_with_invalid_arg(self): bucket_uri = self.CreateBucket() stderr = self.RunGsUtil( ['retention', 'lock', '-a', suri(bucket_uri)], stdin='y', expected_status=1, return_stderr=True) self.assertRegex(stderr, r'Incorrect option\(s\) specified') @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_lock_retention_already_locked(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY, is_locked=True) stderr = self.RunGsUtil( ['retention', 'lock', suri(bucket_uri)], stdin='y', return_stderr=True) self.assertRegex(stderr, r'Retention [Pp]olicy on .* is already locked') @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_increase_retention_locked(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY, is_locked=True) self.RunGsUtil([ 'retention', 'set', '{}s'.format(_SECONDS_IN_DAY + 1), suri(bucket_uri) ]) self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=_SECONDS_IN_DAY + 1, expected_is_locked=True) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_decrease_retention_locked(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY, is_locked=True) stderr = self.RunGsUtil([ 'retention', 'set', '{}s'.format(_SECONDS_IN_DAY - 1), suri(bucket_uri) ], expected_status=1, return_stderr=True) self.assertRegex( stderr, 'Cannot reduce retention duration of a ' 'locked Retention Policy for bucket') @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_clear_locked_retention(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY, is_locked=True) stderr = self.RunGsUtil( ['retention', 'clear', suri(bucket_uri)], expected_status=1, return_stderr=True) self.assertRegex( stderr, r'Bucket .* has a locked Retention Policy which cannot be removed') @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_get_retention_locked(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY, is_locked=True) stdout = self.RunGsUtil( ['-DD', 'retention', 'get', suri(bucket_uri)], return_stdout=True) if self._use_gcloud_storage: self.assertRegex(stdout, r'isLocked\: true') self.assertRegex(stdout, r'retentionPeriod\: \'86400\'') self.assertRegex(stdout, r'effectiveTime\: \'.*\'') else: self.assertRegex(stdout, r'Retention Policy \(LOCKED\):') self.assertRegex(stdout, r'Duration: 1 Day\(s\)') self.assertRegex(stdout, r'Effective Time: .* GMT') actual_retention_policy = self.json_api.GetBucket( bucket_uri.bucket_name, fields=['retentionPolicy']).retentionPolicy if self._use_gcloud_storage: expected_effective_time = datetime.datetime.fromisoformat( re.search(r'effectiveTime\: \'(.*)\'', stdout).group(1)) actual_effective_time = actual_retention_policy.effectiveTime else: expected_effective_time = self._ConvertTimeStringToSeconds( re.search(r'(?<=Time: )[\w,: ]+', stdout).group()) actual_effective_time = self.DateTimeToSeconds( actual_retention_policy.effectiveTime.replace(tzinfo=None)) self.assertEqual(actual_effective_time, expected_effective_time) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_get_retention_unlocked(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY) stdout = self.RunGsUtil( ['retention', 'get', suri(bucket_uri)], return_stdout=True) if self._use_gcloud_storage: # Sometimes the field is absent if isLocked is false. self.assertNotRegex(stdout, r'isLocked \: true') self.assertRegex(stdout, r'retentionPeriod\: \'86400\'') self.assertRegex(stdout, r'effectiveTime\: \'.*\'') else: self.assertRegex(stdout, r'Retention Policy \(UNLOCKED\):') self.assertRegex(stdout, r'Duration: 1 Day\(s\)') self.assertRegex(stdout, r'Effective Time: .* GMT') actual_retention_policy = self.json_api.GetBucket( bucket_uri.bucket_name, fields=['retentionPolicy']).retentionPolicy if self._use_gcloud_storage: expected_effective_time = datetime.datetime.fromisoformat( re.search(r'effectiveTime\: \'(.*)\'', stdout).group(1)) actual_effective_time = actual_retention_policy.effectiveTime else: expected_effective_time = self._ConvertTimeStringToSeconds( re.search(r'(?<=Time: )[\w,: ]+', stdout).group()) actual_effective_time = self.DateTimeToSeconds( actual_retention_policy.effectiveTime.replace(tzinfo=None)) self.assertEqual(actual_effective_time, expected_effective_time) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_get_no_retention(self): bucket_uri = self.CreateBucket() stdout = self.RunGsUtil( ['retention', 'get', suri(bucket_uri)], return_stdout=True) if self._use_gcloud_storage: self.assertRegex(stdout, 'null') else: self.assertRegex(stdout, 'has no Retention Policy') @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_get_invalid_args(self): bucket_uri = self.CreateBucketWithRetentionPolicy( retention_period_in_seconds=_SECONDS_IN_DAY) stderr = self.RunGsUtil( ['retention', 'get', '-a', suri(bucket_uri)], expected_status=1, return_stderr=True) self.assertRegex(stderr, r'Incorrect option\(s\) specified.') @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_set_temporary_hold_invalid_arg(self): object_uri = self.CreateObject() stderr = self.RunGsUtil(['retention', 'temp', 'held', suri(object_uri)], expected_status=1, return_stderr=True) self.assertRegex( stderr, r'Invalid subcommand ".*" for the "retention temp" command') @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_temporary_hold_bucket_with_no_retention(self): bucket_uri = self.CreateBucket() object_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, temporary_hold=None) self.RunGsUtil(['retention', 'temp', 'set', suri(object_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, temporary_hold=True) self.RunGsUtil(['retention', 'temp', 'release', suri(object_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, temporary_hold=False) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_temporary_hold_bucket_with_retention(self): retention_period = 1 bucket_uri = self.CreateBucketWithRetentionPolicy(retention_period) object_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, temporary_hold=None, retention_period=retention_period) self.RunGsUtil(['retention', 'temp', 'set', suri(object_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, temporary_hold=True, retention_period=retention_period) self.RunGsUtil(['retention', 'temp', 'release', suri(object_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, temporary_hold=False, retention_period=retention_period) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_temporary_hold_multiple_sequential(self): bucket_uri = self.CreateBucket() object1_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') object2_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') self.RunGsUtil( ['retention', 'temp', 'set', suri(object1_uri), suri(object2_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object1_uri, temporary_hold=True) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object2_uri, temporary_hold=True) self.RunGsUtil( ['retention', 'temp', 'release', suri(object1_uri), suri(object2_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object1_uri, temporary_hold=False) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object2_uri, temporary_hold=False) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_temporary_hold_multiple_parallel(self): bucket_uri = self.CreateBucket() object1_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') object2_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') self.RunGsUtil([ '-m', 'retention', 'temp', 'set', suri(object1_uri), suri(object2_uri) ]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object1_uri, temporary_hold=True) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object2_uri, temporary_hold=True) self.RunGsUtil([ '-m', 'retention', 'temp', 'release', suri(object1_uri), suri(object2_uri) ]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object1_uri, temporary_hold=False) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object2_uri, temporary_hold=False) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_set_event_based_hold_invalid_arg(self): object_uri = self.CreateObject() stderr = self.RunGsUtil(['retention', 'event', 'rel', suri(object_uri)], expected_status=1, return_stderr=True) self.assertRegex( stderr, r'Invalid subcommand ".*" for the "retention event" command') @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_event_based_hold_bucket_with_no_retention(self): bucket_uri = self.CreateBucket() object_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, event_based_hold=None) self.RunGsUtil(['retention', 'event', 'set', suri(object_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, event_based_hold=True) self.RunGsUtil(['retention', 'event', 'release', suri(object_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, event_based_hold=False) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_event_based_hold_bucket_with_retention(self): retention_period = 1 bucket_uri = self.CreateBucketWithRetentionPolicy(retention_period) object_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, event_based_hold=None, retention_period=retention_period) self.RunGsUtil(['retention', 'event', 'set', suri(object_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, event_based_hold=True, retention_period=None) self.RunGsUtil(['retention', 'event', 'release', suri(object_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object_uri, event_based_hold=False, retention_period=retention_period) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_event_based_hold_multiple_sequential(self): bucket_uri = self.CreateBucket() object1_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') object2_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') self.RunGsUtil( ['retention', 'event', 'set', suri(object1_uri), suri(object2_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object1_uri, event_based_hold=True) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object2_uri, event_based_hold=True) self.RunGsUtil( ['retention', 'event', 'release', suri(object1_uri), suri(object2_uri)]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object1_uri, event_based_hold=False) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object2_uri, event_based_hold=False) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_event_based_hold_multiple_parallel(self): bucket_uri = self.CreateBucket() object1_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') object2_uri = self.CreateObject(bucket_uri=bucket_uri, contents='content') self.RunGsUtil([ '-m', 'retention', 'event', 'set', suri(object1_uri), suri(object2_uri) ]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object1_uri, event_based_hold=True) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object2_uri, event_based_hold=True) self.RunGsUtil([ '-m', 'retention', 'event', 'release', suri(object1_uri), suri(object2_uri) ]) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object1_uri, event_based_hold=False) self._VerifyObjectHoldAndRetentionStatus(bucket_uri, object2_uri, event_based_hold=False) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_default_event_based_hold(self): bucket_uri = self.CreateBucket() self.RunGsUtil(['retention', 'event-default', 'set', suri(bucket_uri)]) self._VerifyDefaultEventBasedHold(bucket_uri, expected_default_event_based_hold=True) self.RunGsUtil(['retention', 'event-default', 'release', suri(bucket_uri)]) self._VerifyDefaultEventBasedHold(bucket_uri, expected_default_event_based_hold=False) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_default_event_based_hold_multiple_sequential(self): bucket1_uri = self.CreateBucket() bucket2_uri = self.CreateBucket() self.RunGsUtil([ 'retention', 'event-default', 'set', suri(bucket1_uri), suri(bucket2_uri) ]) self._VerifyDefaultEventBasedHold(bucket1_uri, expected_default_event_based_hold=True) self._VerifyDefaultEventBasedHold(bucket2_uri, expected_default_event_based_hold=True) @SkipForS3('Retention is not supported for s3 objects') @SkipForXML('Retention is not supported for XML API') def test_default_event_based_hold_multiple_parallel(self): bucket1_uri = self.CreateBucket() bucket2_uri = self.CreateBucket() self.RunGsUtil([ 'retention', 'event-default', 'set', suri(bucket1_uri), suri(bucket2_uri) ]) self._VerifyDefaultEventBasedHold(bucket1_uri, expected_default_event_based_hold=True) self._VerifyDefaultEventBasedHold(bucket2_uri, expected_default_event_based_hold=True) def _VerifyObjectHoldAndRetentionStatus(self, bucket_uri, object_uri, temporary_hold=None, event_based_hold=None, retention_period=None): object_metadata = self.json_api.GetObjectMetadata( bucket_uri.bucket_name, object_uri.object_name, fields=[ 'timeCreated', 'temporaryHold', 'eventBasedHold', 'retentionExpirationTime' ]) if temporary_hold is None: self.assertEqual(object_metadata.temporaryHold, None) else: self.assertEqual(object_metadata.temporaryHold, temporary_hold) if event_based_hold is None: self.assertEqual(object_metadata.eventBasedHold, None) else: self.assertEqual(object_metadata.eventBasedHold, event_based_hold) if retention_period is None: self.assertEqual(object_metadata.retentionExpirationTime, None) elif event_based_hold is False: retention_policy = self.json_api.GetBucket(bucket_uri.bucket_name, fields=['retentionPolicy' ]).retentionPolicy time_delta = datetime.timedelta(0, retention_policy.retentionPeriod) expected_expiration_time = object_metadata.timeCreated + time_delta if event_based_hold is None: self.assertEqual(object_metadata.retentionExpirationTime, expected_expiration_time) else: # since we don't expose the release time of event-based hold we can # only verify that expected_expiration_time is greater than # object-creation-time + retention period # that is because # eventBased-hold's release time > object-creation-time self.assertGreater(object_metadata.retentionExpirationTime, expected_expiration_time) def _VerifyDefaultEventBasedHold(self, bucket_uri, expected_default_event_based_hold=None): actual_default_event_based_hold = self.json_api.GetBucket( bucket_uri.bucket_name, fields=['defaultEventBasedHold']).defaultEventBasedHold if expected_default_event_based_hold is None: self.assertEqual(actual_default_event_based_hold, None) else: self.assertEqual(actual_default_event_based_hold, expected_default_event_based_hold) def _ConvertTimeStringToSeconds(self, time_string): """Converts time in following format to its equivalent timestamp in seconds. Format: '%a, %d %b %Y %H:%M:%S GMT' i.e.: 'Fri, 18 Aug 2017 23:31:39 GMT' Args: time_string: time in string format. Returns: returns equivalent timestamp in seconds of given time. """ converted_time = datetime.datetime.strptime(time_string, '%a, %d %b %Y %H:%M:%S GMT') return self.DateTimeToSeconds(converted_time)