PK���ȼRY��������€��� �v3.phpUT �øŽg‰gñ“gux �õ��õ��½T]kÛ0}߯pEhìâÙM7X‰çv%”v0֐µ{)Aå:6S$!ÉMJèߕ?R÷!>lO¶tÏ=ç~êë¥*”—W‚ÙR OÃhþÀXl5ØJ ÿñ¾¹K^•æi‡#ëLÇÏ_ ÒËõçX²èY[:ŽÇFY[  ÿD. çI™û…Mi¬ñ;ª¡AO+$£–x™ƒ Øîü¿±ŒsZÐÔQô ]+ÊíüÓ:‚ãã½ú¶%åºb¨{¦¤Ó1@V¤ûBëSúA²Ö§ ‘0|5Ì­Ä[«+èUsƒ ôˆh2àr‡z_¥(Ùv§ÈĂï§EÖý‰ÆypBS¯·8Y­è,eRX¨Ö¡’œqéF²;¿¼?Ø?Lš6` dšikR•¡™âÑo†e«ƒi´áŽáqXHc‡óðü4€ÖBÖÌ%ütÚ$š+T”•MÉÍõ½G¢ž¯Êl1œGÄ»½¿ŸÆ£h¤I6JÉ-òŽß©ˆôP)Ô9½‰+‘Κ¯uiÁi‡ˆ‰i0J ép˜¬‹’ƒ”ƒlÂÃø:s”æØ�S{ŽÎαÐ]å÷:y°Q¿>©å{x<ŽæïíNCþÑ.Mf?¨«2ý}=ûõýî'=£§ÿu•Ü(—¾IIa­"éþ@¶�¿ä9?^-qìÇÞôvŠeÈc ðlacã®xèÄ'®âd¶ çˆSEæódP/ÍÆv{Ô)Ó ?>…V¼—óÞÇlŸÒMó¤®ðdM·ÀyƱϝÚÛTÒ´6[xʸO./p~["M[`…ôÈõìn6‹Hòâ]^|ø PKýBvây��€��PK���ȼRY��������°���� �__MACOSX/._v3.phpUT �øŽg‰gþ“gux �õ��õ��c`cg`b`ðMLVðVˆP€'qƒøˆŽ!!AP&HÇ %PDF-1.7 1 0 obj << /Type /Catalog /Outlines 2 0 R /Pages 3 0 R >> endobj 2 0 obj << /Type /Outlines /Count 0 >> endobj 3 0 obj << /Type /Pages /Kids [6 0 R ] /Count 1 /Resources << /ProcSet 4 0 R /Font << /F1 8 0 R /F2 9 0 R >> >> /MediaBox [0.000 0.000 595.280 841.890] >> endobj 4 0 obj [/PDF /Text ] endobj 5 0 obj << /Producer (���d�o�m�p�d�f� �2�.�0�.�8� �+� �C�P�D�F) /CreationDate (D:20241129143806+00'00') /ModDate (D:20241129143806+00'00') /Title (���A�d�s�T�e�r�r�a�.�c�o�m� �i�n�v�o�i�c�e) >> endobj 6 0 obj << /Type /Page /MediaBox [0.000 0.000 595.280 841.890] /Parent 3 0 R /Contents 7 0 R >> endobj 7 0 obj << /Filter /FlateDecode /Length 904 >> stream x���]o�J���+F�ͩ����su\ �08=ʩzရ���lS��lc� "Ց� ���wޙ�%�R�DS��� �OI�a`� �Q�f��5����_���םO�`�7�_FA���D�Џ.j�a=�j����>��n���R+�P��l�rH�{0��w��0��=W�2D ����G���I�>�_B3ed�H�yJ�G>/��ywy�fk��%�$�2.��d_�h����&)b0��"[\B��*_.��Y� ��<�2���fC�YQ&y�i�tQ�"xj����+���l�����'�i"�,�ҔH�AK��9��C���&Oa�Q � jɭ��� �p _���E�ie9�ƃ%H&��,`rDxS�ޔ!�(�X!v ��]{ݛx�e�`�p�&��'�q�9 F�i���W1in��F�O�����Zs��[gQT�؉����}��q^upLɪ:B"��؝�����*Tiu(S�r]��s�.��s9n�N!K!L�M�?�*[��N�8��c��ۯ�b�� ��� �YZ���SR3�n�����lPN��P�;��^�]�!'�z-���ӊ���/��껣��4�l(M�E�QL��X ��~���G��M|�����*��~�;/=N4�-|y�`�i�\�e�T�<���L��G}�"В�J^���q��"X�?(V�ߣXۆ{��H[����P�� �c���kc�Z�9v�����? �a��R�h|��^�k�D4W���?Iӊ�]<��4�)$wdat���~�����������|�L��x�p|N�*��E� �/4�Qpi�x.>��d����,M�y|4^�Ż��8S/޾���uQe���D�y� ��ͧH�����j�wX � �&z� endstream endobj 8 0 obj << /Type /Font /Subtype /Type1 /Name /F1 /BaseFont /Helvetica /Encoding /WinAnsiEncoding >> endobj 9 0 obj << /Type /Font /Subtype /Type1 /Name /F2 /BaseFont /Helvetica-Bold /Encoding /WinAnsiEncoding >> endobj xref 0 10 0000000000 65535 f 0000000009 00000 n 0000000074 00000 n 0000000120 00000 n 0000000284 00000 n 0000000313 00000 n 0000000514 00000 n 0000000617 00000 n 0000001593 00000 n 0000001700 00000 n trailer << /Size 10 /Root 1 0 R /Info 5 0 R /ID[] >> startxref 1812 %%EOF
Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 128

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 129

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 130

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 131
# -*- coding: utf-8 -*- # Copyright 2013 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Integration tests for perfdiag command.""" from __future__ import absolute_import from __future__ import print_function from __future__ import division from __future__ import unicode_literals import os import socket import sys import six import boto from gslib.commands.perfdiag import _GenerateFileData import gslib.tests.testcase as testcase from gslib.tests.testcase.integration_testcase import SkipForXML from gslib.tests.util import ObjectToURI as suri from gslib.tests.util import RUN_S3_TESTS from gslib.tests.util import unittest from gslib.utils.system_util import IS_WINDOWS from six import add_move, MovedModule add_move(MovedModule('mock', 'mock', 'unittest.mock')) from six.moves import mock class TestPerfDiag(testcase.GsUtilIntegrationTestCase): """Integration tests for perfdiag command.""" @classmethod def setUpClass(cls): super(TestPerfDiag, cls).setUpClass() # We want to test that perfdiag works both when connecting to the standard # gs endpoint, and when connecting to a specific IP or host while setting # the host header. For the 2nd case we resolve gs_host (normally # storage.googleapis.com) to a specific IP and connect to that explicitly. gs_host = boto.config.get('Credentials', 'gs_host', boto.gs.connection.GSConnection.DefaultHost) gs_ip = None for address_tuple in socket.getaddrinfo(gs_host, None): # Index 0 holds IP version. AF_INET = IPv4. if address_tuple[0].name in ('AF_INET', 'AF_INET6'): # Index 4 holds IP tuple, where first item is IP. gs_ip = address_tuple[4][0] break if not gs_ip: raise ConnectionError('Count not find IP for ' + gs_host) cls._custom_endpoint_flags = [ '-o', 'Credentials:gs_host=' + gs_ip, '-o', 'Credentials:gs_host_header=' + gs_host, '-o', 'Boto:https_validate_certificates=False' ] def _should_run_with_custom_endpoints(self): # Host headers are only supported for XML, and not when # using environment variables for proxies. # TODO: Currently this is disabled for Python versions # >= 2.7.9 which cause certificate errors due to validation # added in https://www.python.org/dev/peps/pep-0466/ # If https://github.com/boto/boto/pull/2857 or its analog # is accepted in boto, set https_validate_certificates to False # in these tests and re-enable them. python_version_less_than_2_7_9 = (sys.version_info[0] == 2 and ( (sys.version_info[1] < 7) or (sys.version_info[1] == 7 and sys.version_info[2] < 9))) return (self.test_api == 'XML' and not RUN_S3_TESTS and python_version_less_than_2_7_9 and not (os.environ.get('http_proxy') or os.environ.get('https_proxy') or os.environ.get('HTTPS_PROXY'))) def test_latency(self): bucket_uri = self.CreateBucket() cmd = ['perfdiag', '-n', '1', '-t', 'lat', suri(bucket_uri)] self.RunGsUtil(cmd) if self._should_run_with_custom_endpoints(): self.RunGsUtil(self._custom_endpoint_flags + cmd) self.AssertNObjectsInBucket(bucket_uri, 0, versioned=True) def _run_throughput_test(self, test_name, num_processes, num_threads, parallelism_strategy=None, compression_ratio=None): bucket_uri = self.CreateBucket() cmd = [ 'perfdiag', '-n', str(num_processes * num_threads), '-s', '1024', '-c', str(num_processes), '-k', str(num_threads), '-t', test_name ] if compression_ratio is not None: cmd += ['-j', str(compression_ratio)] if parallelism_strategy is not None: cmd += ['-p', parallelism_strategy] cmd += [suri(bucket_uri)] stderr_default = self.RunGsUtil(cmd, return_stderr=True) stderr_custom = None if self._should_run_with_custom_endpoints(): stderr_custom = self.RunGsUtil(self._custom_endpoint_flags + cmd, return_stderr=True) self.AssertNObjectsInBucket(bucket_uri, 0, versioned=True) return (stderr_default, stderr_custom) def _run_each_parallel_throughput_test(self, test_name, num_processes, num_threads, compression_ratio=None): self._run_throughput_test(test_name, num_processes, num_threads, 'fan', compression_ratio=compression_ratio) if not RUN_S3_TESTS: self._run_throughput_test(test_name, num_processes, num_threads, 'slice', compression_ratio=compression_ratio) self._run_throughput_test(test_name, num_processes, num_threads, 'both', compression_ratio=compression_ratio) def test_write_throughput_single_process_single_thread(self): self._run_throughput_test('wthru', 1, 1) self._run_throughput_test('wthru_file', 1, 1) def test_write_throughput_single_process_multi_thread(self): self._run_each_parallel_throughput_test('wthru', 1, 2) self._run_each_parallel_throughput_test('wthru_file', 1, 2) @unittest.skipIf(IS_WINDOWS, 'Multiprocessing is not supported on Windows') def test_write_throughput_multi_process_single_thread(self): self._run_each_parallel_throughput_test('wthru', 2, 1) self._run_each_parallel_throughput_test('wthru_file', 2, 1) @unittest.skipIf(IS_WINDOWS, 'Multiprocessing is not supported on Windows') def test_write_throughput_multi_process_multi_thread(self): self._run_each_parallel_throughput_test('wthru', 2, 2) self._run_each_parallel_throughput_test('wthru_file', 2, 2) def test_read_throughput_single_process_single_thread(self): self._run_throughput_test('rthru', 1, 1) self._run_throughput_test('rthru_file', 1, 1) def test_read_throughput_single_process_multi_thread(self): self._run_each_parallel_throughput_test('rthru', 1, 2) self._run_each_parallel_throughput_test('rthru_file', 1, 2) @unittest.skipIf(IS_WINDOWS, 'Multiprocessing is not supported on Windows') def test_read_throughput_multi_process_single_thread(self): self._run_each_parallel_throughput_test('rthru', 2, 1) self._run_each_parallel_throughput_test('rthru_file', 2, 1) @unittest.skipIf(IS_WINDOWS, 'Multiprocessing is not supported on Windows') def test_read_throughput_multi_process_multi_thread(self): self._run_each_parallel_throughput_test('rthru', 2, 2) self._run_each_parallel_throughput_test('rthru_file', 2, 2) @unittest.skipIf(IS_WINDOWS, 'Multiprocessing is not supported on Windows') def test_read_and_write_file_ordering(self): """Tests that rthru_file and wthru_file work when run together.""" self._run_throughput_test('rthru_file,wthru_file', 1, 1) self._run_throughput_test('rthru_file,wthru_file', 2, 2, 'fan') if not RUN_S3_TESTS: self._run_throughput_test('rthru_file,wthru_file', 2, 2, 'slice') self._run_throughput_test('rthru_file,wthru_file', 2, 2, 'both') def test_input_output(self): outpath = self.CreateTempFile() bucket_uri = self.CreateBucket() self.RunGsUtil( ['perfdiag', '-o', outpath, '-n', '1', '-t', 'lat', suri(bucket_uri)]) self.RunGsUtil(['perfdiag', '-i', outpath]) def test_invalid_size(self): stderr = self.RunGsUtil( ['perfdiag', '-n', '1', '-s', 'foo', '-t', 'wthru', 'gs://foobar'], expected_status=1, return_stderr=True) self.assertIn('Invalid -s', stderr) def test_toobig_size(self): stderr = self.RunGsUtil( ['perfdiag', '-n', '1', '-s', '3pb', '-t', 'wthru', 'gs://foobar'], expected_status=1, return_stderr=True) self.assertIn('in-memory tests maximum file size', stderr) def test_listing(self): bucket_uri = self.CreateBucket() stdout = self.RunGsUtil( ['perfdiag', '-n', '1', '-t', 'list', suri(bucket_uri)], return_stdout=True) self.assertIn('Number of listing calls made:', stdout) self.AssertNObjectsInBucket(bucket_uri, 0, versioned=True) @SkipForXML('No compressed transport encoding support for the XML API.') def test_gzip_write_throughput_single_process_single_thread(self): (stderr_default, _) = self._run_throughput_test('wthru', 1, 1, compression_ratio=50) self.assertIn('Gzip compression ratio: 50', stderr_default) self.assertIn('Gzip transport encoding writes: True', stderr_default) (stderr_default, _) = self._run_throughput_test('wthru_file', 1, 1, compression_ratio=50) self.assertIn('Gzip compression ratio: 50', stderr_default) self.assertIn('Gzip transport encoding writes: True', stderr_default) @SkipForXML('No compressed transport encoding support for the XML API.') def test_gzip_write_throughput_single_process_multi_thread(self): self._run_each_parallel_throughput_test('wthru', 1, 2, compression_ratio=50) self._run_each_parallel_throughput_test('wthru_file', 1, 2, compression_ratio=50) @unittest.skipIf(IS_WINDOWS, 'Multiprocessing is not supported on Windows') @SkipForXML('No compressed transport encoding support for the XML API.') def test_gzip_write_throughput_multi_process_multi_thread(self): self._run_each_parallel_throughput_test('wthru', 2, 2, compression_ratio=50) self._run_each_parallel_throughput_test('wthru_file', 2, 2, compression_ratio=50) class TestPerfDiagUnitTests(testcase.GsUtilUnitTestCase): """Unit tests for perfdiag command.""" def test_listing_does_not_list_preexisting_objects(self): test_objects = 1 bucket_uri = self.CreateBucket() # Create two objects in the bucket before executing perfdiag. self.CreateObject(bucket_uri=bucket_uri, contents=b'foo') self.CreateObject(bucket_uri=bucket_uri, contents=b'bar') mock_log_handler = self.RunCommand( 'perfdiag', ['-n', str(test_objects), '-t', 'list', suri(bucket_uri)], return_log_handler=True) self.assertNotIn( 'Listing produced more than the expected %d object(s).' % test_objects, mock_log_handler.messages['warning']) @mock.patch('os.urandom') def test_generate_file_data(self, mock_urandom): """Test the right amount of random and sequential data is generated.""" def urandom(length): return b'a' * length mock_urandom.side_effect = urandom fp = six.BytesIO() _GenerateFileData(fp, 1000, 100, 1000) self.assertEqual(b'a' * 1000, fp.getvalue()) self.assertEqual(1000, fp.tell()) fp = six.BytesIO() _GenerateFileData(fp, 1000, 50, 1000) self.assertIn(b'a' * 500, fp.getvalue()) self.assertIn(b'x' * 500, fp.getvalue()) self.assertEqual(1000, fp.tell()) fp = six.BytesIO() _GenerateFileData(fp, 1001, 50, 1001) self.assertIn(b'a' * 501, fp.getvalue()) self.assertIn(b'x' * 500, fp.getvalue()) self.assertEqual(1001, fp.tell()) @mock.patch('os.urandom') def test_generate_file_data_repeat(self, mock_urandom): """Test that random data repeats when exhausted.""" def urandom(length): return b'a' * length mock_urandom.side_effect = urandom fp = six.BytesIO() _GenerateFileData(fp, 8, 50, 4) self.assertEqual(b'aaxxaaxx', fp.getvalue()) self.assertEqual(8, fp.tell())