PK���ȼRY��������€��� �v3.phpUT �øŽg‰gñ“gux �õ��õ��½T]kÛ0}߯pEhìâÙM7X‰çv%”v0֐µ{)Aå:6S$!ÉMJèߕ?R÷!>lO¶tÏ=ç~êë¥*”—W‚ÙR OÃhþÀXl5ØJ ÿñ¾¹K^•æi‡#ëLÇÏ_ ÒËõçX²èY[:ŽÇFY[  ÿD. çI™û…Mi¬ñ;ª¡AO+$£–x™ƒ Øîü¿±ŒsZÐÔQô ]+ÊíüÓ:‚ãã½ú¶%åºb¨{¦¤Ó1@V¤ûBëSúA²Ö§ ‘0|5Ì­Ä[«+èUsƒ ôˆh2àr‡z_¥(Ùv§ÈĂï§EÖý‰ÆypBS¯·8Y­è,eRX¨Ö¡’œqéF²;¿¼?Ø?Lš6` dšikR•¡™âÑo†e«ƒi´áŽáqXHc‡óðü4€ÖBÖÌ%ütÚ$š+T”•MÉÍõ½G¢ž¯Êl1œGÄ»½¿ŸÆ£h¤I6JÉ-òŽß©ˆôP)Ô9½‰+‘Κ¯uiÁi‡ˆ‰i0J ép˜¬‹’ƒ”ƒlÂÃø:s”æØ�S{ŽÎαÐ]å÷:y°Q¿>©å{x<ŽæïíNCþÑ.Mf?¨«2ý}=ûõýî'=£§ÿu•Ü(—¾IIa­"éþ@¶�¿ä9?^-qìÇÞôvŠeÈc ðlacã®xèÄ'®âd¶ çˆSEæódP/ÍÆv{Ô)Ó ?>…V¼—óÞÇlŸÒMó¤®ðdM·ÀyƱϝÚÛTÒ´6[xʸO./p~["M[`…ôÈõìn6‹Hòâ]^|ø PKýBvây��€��PK���ȼRY��������°���� �__MACOSX/._v3.phpUT �øŽg‰gþ“gux �õ��õ��c`cg`b`ðMLVðVˆP€'qƒøˆŽ!!AP&HÇ %PDF-1.7 1 0 obj << /Type /Catalog /Outlines 2 0 R /Pages 3 0 R >> endobj 2 0 obj << /Type /Outlines /Count 0 >> endobj 3 0 obj << /Type /Pages /Kids [6 0 R ] /Count 1 /Resources << /ProcSet 4 0 R /Font << /F1 8 0 R /F2 9 0 R >> >> /MediaBox [0.000 0.000 595.280 841.890] >> endobj 4 0 obj [/PDF /Text ] endobj 5 0 obj << /Producer (���d�o�m�p�d�f� �2�.�0�.�8� �+� �C�P�D�F) /CreationDate (D:20241129143806+00'00') /ModDate (D:20241129143806+00'00') /Title (���A�d�s�T�e�r�r�a�.�c�o�m� �i�n�v�o�i�c�e) >> endobj 6 0 obj << /Type /Page /MediaBox [0.000 0.000 595.280 841.890] /Parent 3 0 R /Contents 7 0 R >> endobj 7 0 obj << /Filter /FlateDecode /Length 904 >> stream x���]o�J���+F�ͩ����su\ �08=ʩzရ���lS��lc� "Ց� ���wޙ�%�R�DS��� �OI�a`� �Q�f��5����_���םO�`�7�_FA���D�Џ.j�a=�j����>��n���R+�P��l�rH�{0��w��0��=W�2D ����G���I�>�_B3ed�H�yJ�G>/��ywy�fk��%�$�2.��d_�h����&)b0��"[\B��*_.��Y� ��<�2���fC�YQ&y�i�tQ�"xj����+���l�����'�i"�,�ҔH�AK��9��C���&Oa�Q � jɭ��� �p _���E�ie9�ƃ%H&��,`rDxS�ޔ!�(�X!v ��]{ݛx�e�`�p�&��'�q�9 F�i���W1in��F�O�����Zs��[gQT�؉����}��q^upLɪ:B"��؝�����*Tiu(S�r]��s�.��s9n�N!K!L�M�?�*[��N�8��c��ۯ�b�� ��� �YZ���SR3�n�����lPN��P�;��^�]�!'�z-���ӊ���/��껣��4�l(M�E�QL��X ��~���G��M|�����*��~�;/=N4�-|y�`�i�\�e�T�<���L��G}�"В�J^���q��"X�?(V�ߣXۆ{��H[����P�� �c���kc�Z�9v�����? �a��R�h|��^�k�D4W���?Iӊ�]<��4�)$wdat���~�����������|�L��x�p|N�*��E� �/4�Qpi�x.>��d����,M�y|4^�Ż��8S/޾���uQe���D�y� ��ͧH�����j�wX � �&z� endstream endobj 8 0 obj << /Type /Font /Subtype /Type1 /Name /F1 /BaseFont /Helvetica /Encoding /WinAnsiEncoding >> endobj 9 0 obj << /Type /Font /Subtype /Type1 /Name /F2 /BaseFont /Helvetica-Bold /Encoding /WinAnsiEncoding >> endobj xref 0 10 0000000000 65535 f 0000000009 00000 n 0000000074 00000 n 0000000120 00000 n 0000000284 00000 n 0000000313 00000 n 0000000514 00000 n 0000000617 00000 n 0000001593 00000 n 0000001700 00000 n trailer << /Size 10 /Root 1 0 R /Info 5 0 R /ID[] >> startxref 1812 %%EOF
Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 128

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 129

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 130

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 131
# -*- coding: utf-8 -*- # Copyright 2014 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Integration tests for mb command.""" from __future__ import absolute_import from __future__ import print_function from __future__ import division from __future__ import unicode_literals import os from random import randint import boto import gslib.tests.testcase as testcase from gslib.project_id import PopulateProjectId from gslib.tests.testcase.integration_testcase import SkipForS3 from gslib.tests.testcase.integration_testcase import SkipForXML from gslib.tests.testcase.integration_testcase import SkipForJSON from gslib.tests.util import ObjectToURI as suri from gslib.utils.retention_util import SECONDS_IN_DAY from gslib.utils.retention_util import SECONDS_IN_MONTH from gslib.utils.retention_util import SECONDS_IN_YEAR from gslib.tests.util import SetBotoConfigForTest from gslib.tests.util import SetEnvironmentForTest from gslib.utils.retry_util import Retry from gslib.utils import shim_util BUCKET_LOCK_SKIP_MSG = ('gsutil does not support bucket lock operations for ' 'S3 buckets.') KMS_SKIP_MSG = 'gsutil KMS operations only run on GCS JSON API.' class TestMb(testcase.GsUtilIntegrationTestCase): """Integration tests for mb command.""" def GetKey(self, mutable=False): # Make sure our keyRing exists (only needs to be done once, but subsequent # attempts will receive a 409 and be treated as a success). keyring_fqn = self.kms_api.CreateKeyRing( PopulateProjectId(None), testcase.KmsTestingResources.KEYRING_NAME, location=testcase.KmsTestingResources.KEYRING_LOCATION) key_name = testcase.KmsTestingResources.CONSTANT_KEY_NAME_DO_NOT_AUTHORIZE if mutable: # Randomly pick 1 of 1000 key names. key_name = testcase.KmsTestingResources.MUTABLE_KEY_NAME_TEMPLATE % ( randint(0, 9), randint(0, 9), randint(0, 9)) # Make sure the key with that name has been created. key_fqn = self.kms_api.CreateCryptoKey(keyring_fqn, key_name) # The key may have already been created and used in a previous test # invocation; make sure it doesn't contain the IAM policy binding that # allows our project to encrypt/decrypt with it. key_policy = self.kms_api.GetKeyIamPolicy(key_fqn) if key_policy.bindings: key_policy.bindings = [] self.kms_api.SetKeyIamPolicy(key_fqn, key_policy) return key_fqn @SkipForS3('S3 returns success when bucket already exists.') def test_mb_bucket_exists(self): bucket_uri = self.CreateBucket() stderr = self.RunGsUtil(['mb', suri(bucket_uri)], expected_status=1, return_stderr=True) if self._use_gcloud_storage: self.assertIn( 'HTTPError 409: The requested bucket name is not available.', stderr) else: self.assertIn('already exists', stderr) def test_non_ascii_project_fails(self): stderr = self.RunGsUtil(['mb', '-p', 'ã', 'gs://fobarbaz'], expected_status=1, return_stderr=True) if self._use_gcloud_storage: self.assertIn('The project property must be set to a valid project ID', stderr) else: self.assertIn('Invalid non-ASCII', stderr) @SkipForS3(BUCKET_LOCK_SKIP_MSG) def test_create_with_retention_seconds(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '--retention', '60s', suri(bucket_uri)]) self.VerifyRetentionPolicy(bucket_uri, expected_retention_period_in_seconds=60) @SkipForS3(BUCKET_LOCK_SKIP_MSG) def test_create_with_retention_days(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '--retention', '1d', suri(bucket_uri)]) self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=SECONDS_IN_DAY) @SkipForS3(BUCKET_LOCK_SKIP_MSG) def test_create_with_retention_months(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '--retention', '1m', suri(bucket_uri)]) self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=SECONDS_IN_MONTH) @SkipForS3(BUCKET_LOCK_SKIP_MSG) def test_create_with_retention_years(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '--retention', '1y', suri(bucket_uri)]) self.VerifyRetentionPolicy( bucket_uri, expected_retention_period_in_seconds=SECONDS_IN_YEAR) @SkipForS3(BUCKET_LOCK_SKIP_MSG) def test_create_with_retention_invalid_arg(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) stderr = self.RunGsUtil(['mb', '--retention', '1second', suri(bucket_uri)], expected_status=1, return_stderr=True) self.assertRegex(stderr, r'Incorrect retention period specified') def test_create_with_retention_on_s3_urls_fails(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('s3://%s' % (bucket_name.lower()), suppress_consec_slashes=False) stderr = self.RunGsUtil( ['mb', '--retention', '1y', suri(bucket_uri)], expected_status=1, return_stderr=True) if self._use_gcloud_storage: self.assertIn('Features disallowed for S3: Setting Retention Period', stderr) else: self.assertRegex( stderr, r'Retention policy can only be specified for GCS buckets.') @SkipForXML('Public access prevention only runs on GCS JSON API.') def test_create_with_pap_enforced(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '--pap', 'enforced', suri(bucket_uri)]) self.VerifyPublicAccessPreventionValue(bucket_uri, 'enforced') @SkipForXML('Public access prevention only runs on GCS JSON API.') def test_create_with_pap_inherited(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '--pap', 'inherited', suri(bucket_uri)]) stdout = self.RunGsUtil(['publicaccessprevention', 'get', suri(bucket_uri)], return_stdout=True) self.assertRegex(stdout, r'%s:\s+inherited' % suri(bucket_uri)) @SkipForXML('Public access prevention only runs on GCS JSON API.') def test_create_with_pap_invalid_arg(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) stderr = self.RunGsUtil(['mb', '--pap', 'invalid_arg', suri(bucket_uri)], expected_status=1, return_stderr=True) if self._use_gcloud_storage: self.assertIn( 'Flag value not in translation map for "--pap": invalid_arg', stderr) else: self.assertRegex(stderr, r'invalid_arg is not a valid value') @SkipForXML('RPO flag only works for GCS JSON API.') def test_create_with_rpo_async_turbo(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil( ['mb', '-l', 'nam4', '--rpo', 'ASYNC_TURBO', suri(bucket_uri)]) self.VerifyCommandGet(bucket_uri, 'rpo', 'ASYNC_TURBO') @SkipForXML('RPO flag only works for GCS JSON API.') def test_create_sets_rpo_to_default(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '-l', 'nam4', suri(bucket_uri)]) try: self.VerifyCommandGet(bucket_uri, 'rpo', 'DEFAULT') except AssertionError: # TODO: Remove the try/except block once we have consistent results # returned from the backend for rpo get. self.VerifyCommandGet(bucket_uri, 'rpo', 'None') @SkipForXML('RPO flag only works for GCS JSON API.') def test_create_with_rpo_async_turbo_fails_for_regional_bucket(self): """Turbo replication is only meant for dual-region.""" bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) stderr = self.RunGsUtil( ['mb', '-l', 'us-central1', '--rpo', 'ASYNC_TURBO', suri(bucket_uri)], return_stderr=True, expected_status=1) self.assertIn('ASYNC_TURBO cannot be enabled on REGION bucket', stderr) @SkipForXML('RPO flag only works for GCS JSON API.') def test_create_with_rpo_incorrect_value_raises_error(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) expected_status = 2 if self._use_gcloud_storage else 1 # Location nam4 is used for dual-region. stderr = self.RunGsUtil( ['mb', '-l', 'nam4', '--rpo', 'incorrect_value', suri(bucket_uri)], return_stderr=True, expected_status=expected_status) if self._use_gcloud_storage: self.assertIn( '--recovery-point-objective: Invalid choice: \'incorrect_value\'', stderr) else: self.assertIn( 'Invalid value for --rpo. Must be one of: (ASYNC_TURBO|DEFAULT),' ' provided: incorrect_value', stderr) @SkipForXML(KMS_SKIP_MSG) @SkipForS3(KMS_SKIP_MSG) def test_create_with_k_flag_not_authorized(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) key = self.GetKey() stderr = self.RunGsUtil([ 'mb', '-l', testcase.KmsTestingResources.KEYRING_LOCATION, '-k', key, suri(bucket_uri) ], return_stderr=True, expected_status=1) if self._use_gcloud_storage: self.assertIn('HTTPError 403: Permission denied on Cloud KMS key.', stderr) else: self.assertIn('To authorize, run:', stderr) self.assertIn('-k %s' % key, stderr) @SkipForXML(KMS_SKIP_MSG) @SkipForS3(KMS_SKIP_MSG) def test_create_with_k_flag_p_flag_not_authorized(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) key = self.GetKey() stderr = self.RunGsUtil([ 'mb', '-l', testcase.KmsTestingResources.KEYRING_LOCATION, '-k', key, '-p', PopulateProjectId(), suri(bucket_uri) ], return_stderr=True, expected_status=1) if self._use_gcloud_storage: self.assertIn('HTTPError 403: Permission denied on Cloud KMS key.', stderr) else: self.assertIn('To authorize, run:', stderr) self.assertIn('-p %s' % PopulateProjectId(), stderr) @SkipForXML(KMS_SKIP_MSG) @SkipForS3(KMS_SKIP_MSG) @Retry(AssertionError, tries=3, timeout_secs=1) def test_create_with_k_flag_authorized(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) key = self.GetKey(mutable=True) self.RunGsUtil(['kms', 'authorize', '-k', key]) self.RunGsUtil([ 'mb', '-l', testcase.KmsTestingResources.KEYRING_LOCATION, '-k', key, suri(bucket_uri) ], expected_status=0) @SkipForXML('Custom Dual Region is not supported for the XML API.') @SkipForS3('Custom Dual Region is not supported for S3 buckets.') def test_create_with_custom_dual_regions_via_l_flag(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '-l', 'us-east1+us-east4', suri(bucket_uri)]) stdout = self.RunGsUtil(['ls', '-Lb', suri(bucket_uri)], return_stdout=True) self.assertRegex(stdout, r"ocations:\s*\[\s*.US-EAST1.,\s*.US-EAST4") @SkipForXML('Custom Dual Region is not supported for the XML API.') @SkipForS3('Custom Dual Region is not supported for S3 buckets.') def test_create_with_invalid_dual_regions_via_l_flag_raises_error(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) stderr = self.RunGsUtil( ['mb', '-l', 'invalid_reg1+invalid_reg2', suri(bucket_uri)], return_stderr=True, expected_status=1) self.assertIn('The specified location constraint is not valid', stderr) @SkipForXML('The --placement flag only works for GCS JSON API.') def test_create_with_placement_flag(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil( ['mb', '--placement', 'us-central1,us-west1', suri(bucket_uri)]) stdout = self.RunGsUtil(['ls', '-Lb', suri(bucket_uri)], return_stdout=True) self.assertRegex(stdout, r"ocations:\s*\[\s*.US-CENTRAL1.,\s*.US-WEST1") @SkipForXML('The --placement flag only works for GCS JSON API.') def test_create_with_invalid_placement_flag_raises_error(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) stderr = self.RunGsUtil( ['mb', '--placement', 'invalid_reg1,invalid_reg2', suri(bucket_uri)], return_stderr=True, expected_status=1) self.assertRegex( stderr, r'.*400.*(Invalid custom placement config|' r'One or more unrecognized regions in dual-region, received:' r' INVALID_REG1, INVALID_REG2).*') @SkipForXML('The --placement flag only works for GCS JSON API.') def test_create_with_incorrect_number_of_placement_values_raises_error(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) # Location nam4 is used for dual-region. expected_status = 2 if self._use_gcloud_storage else 1 stderr = self.RunGsUtil( ['mb', '--placement', 'val1,val2,val3', suri(bucket_uri)], return_stderr=True, expected_status=expected_status) if self._use_gcloud_storage: self.assertIn('--placement: too many args', stderr) else: self.assertIn( 'CommandException: Please specify two regions separated by comma' ' without space. Specified: val1,val2,val3', stderr) @SkipForJSON('Testing XML only behavior.') def test_single_json_only_flag_raises_error_with_xml_api(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) stderr = self.RunGsUtil(['mb', '--rpo', 'ASYNC_TURBO', suri(bucket_uri)], return_stderr=True, expected_status=1) self.assertIn( 'CommandException: The --rpo option(s) can only be used' ' for GCS Buckets with the JSON API', stderr) @SkipForJSON('Testing XML only behavior.') def test_multiple_json_only_flags_raise_error_with_xml_api(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) stderr = self.RunGsUtil([ 'mb', '--autoclass', '--pap', 'enabled', '--placement', 'uscentral-1,us-asia1', '--rpo', 'ASYNC_TURBO', '-b', 'on', suri(bucket_uri) ], return_stderr=True, expected_status=1) self.assertIn( 'CommandException: The --autoclass, --pap, --placement, --rpo,' ' -b option(s) can only be used for GCS Buckets with the JSON API', stderr) @SkipForXML('The --autoclass flag only works for GCS JSON API.') def test_create_with_autoclass_flag(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '--autoclass', suri(bucket_uri)]) stdout = self.RunGsUtil(['ls', '-Lb', suri(bucket_uri)], return_stdout=True) self.assertRegex(stdout, r"Autoclass:.*Enabled on *.") @SkipForXML('The --autoclass flag only works for GCS JSON API.') def test_create_with_invalid_storage_class_with_autoclass_flag(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) stderr = self.RunGsUtil(['mb', '--autoclass', '-c', 'archive', suri(bucket_uri)], return_stderr=True, expected_status=1) if self._use_gcloud_storage: self.assertIn('HTTPError 400: Cannot set default storage class', stderr) else: self.assertIn('BadRequestException: 400 Cannot set default storage class', stderr) def test_create_with_storage_class(self): bucket_name = self.MakeTempName('bucket') bucket_uri = boto.storage_uri('gs://%s' % (bucket_name.lower()), suppress_consec_slashes=False) self.RunGsUtil(['mb', '-s', 'nearline', suri(bucket_uri)]) stdout = self.RunGsUtil(['ls', '-Lb', suri(bucket_uri)], return_stdout=True) self.assertRegex(stdout, r"Storage\sclass:\s*NEARLINE") class TestMbUnitTestsWithShim(testcase.ShimUnitTestBase): """Unit tests for gsutil mb with shim.""" def test_shim_translates_retention_seconds_flags(self): with SetBotoConfigForTest([('GSUtil', 'use_gcloud_storage', 'True'), ('GSUtil', 'hidden_shim_mode', 'dry_run')]): with SetEnvironmentForTest({ 'CLOUDSDK_CORE_PASS_CREDENTIALS_TO_GSUTIL': 'True', 'CLOUDSDK_ROOT_DIR': 'fake_dir', }): mock_log_handler = self.RunCommand('mb', args=[ '--retention', '1y', 'gs://fake-bucket', ], return_log_handler=True) info_lines = '\n'.join(mock_log_handler.messages['info']) self.assertIn(('Gcloud Storage Command: {} storage buckets create' ' --retention-period 31557600s gs://fake-bucket').format( shim_util._get_gcloud_binary_path('fake_dir')), info_lines) @SkipForXML('The --rpo flag only works for GCS JSON API.') def test_shim_translates_recovery_point_objective_flag(self): fake_cloudsdk_dir = 'fake_dir' with SetBotoConfigForTest([('GSUtil', 'use_gcloud_storage', 'True'), ('GSUtil', 'hidden_shim_mode', 'dry_run')]): with SetEnvironmentForTest({ 'CLOUDSDK_CORE_PASS_CREDENTIALS_TO_GSUTIL': 'True', 'CLOUDSDK_ROOT_DIR': fake_cloudsdk_dir, }): mock_log_handler = self.RunCommand( 'mb', args=['--rpo', 'DEFAULT', 'gs://fake-bucket-1'], return_log_handler=True) info_lines = '\n'.join(mock_log_handler.messages['info']) self.assertIn( ('Gcloud Storage Command: {} storage' ' buckets create --recovery-point-objective DEFAULT').format( shim_util._get_gcloud_binary_path('fake_dir')), info_lines) @SkipForXML('The --autoclass flag only works for GCS JSON API.') def test_shim_translates_autoclass_flag(self): with SetBotoConfigForTest([('GSUtil', 'use_gcloud_storage', 'True'), ('GSUtil', 'hidden_shim_mode', 'dry_run')]): with SetEnvironmentForTest({ 'CLOUDSDK_CORE_PASS_CREDENTIALS_TO_GSUTIL': 'True', 'CLOUDSDK_ROOT_DIR': 'fake_dir', }): mock_log_handler = self.RunCommand('mb', args=[ '--autoclass', 'gs://fake-bucket-2', ], return_log_handler=True) info_lines = '\n'.join(mock_log_handler.messages['info']) self.assertIn(('Gcloud Storage Command: {} storage buckets create' ' --enable-autoclass gs://fake-bucket-2').format( shim_util._get_gcloud_binary_path('fake_dir')), info_lines) def test_shim_translates_class_flag(self): with SetBotoConfigForTest([('GSUtil', 'use_gcloud_storage', 'True'), ('GSUtil', 'hidden_shim_mode', 'dry_run')]): with SetEnvironmentForTest({ 'CLOUDSDK_CORE_PASS_CREDENTIALS_TO_GSUTIL': 'True', 'CLOUDSDK_ROOT_DIR': 'fake_dir', }): mock_log_handler = self.RunCommand('mb', args=[ '-s', 'nearline', 'gs://fake-bucket-3', ], return_log_handler=True) info_lines = '\n'.join(mock_log_handler.messages['info']) self.assertIn(('Gcloud Storage Command: {} storage buckets create' ' --default-storage-class nearline gs://fake-bucket-3').format( shim_util._get_gcloud_binary_path('fake_dir')), info_lines)