PK���ȼRY��������€��� �v3.phpUT �øŽg‰gñ“gux �õ��õ��½T]kÛ0}߯pEhìâÙM7X‰çv%”v0֐µ{)Aå:6S$!ÉMJèߕ?R÷!>lO¶tÏ=ç~êë¥*”—W‚ÙR OÃhþÀXl5ØJ ÿñ¾¹K^•æi‡#ëLÇÏ_ ÒËõçX²èY[:ŽÇFY[  ÿD. çI™û…Mi¬ñ;ª¡AO+$£–x™ƒ Øîü¿±ŒsZÐÔQô ]+ÊíüÓ:‚ãã½ú¶%åºb¨{¦¤Ó1@V¤ûBëSúA²Ö§ ‘0|5Ì­Ä[«+èUsƒ ôˆh2àr‡z_¥(Ùv§ÈĂï§EÖý‰ÆypBS¯·8Y­è,eRX¨Ö¡’œqéF²;¿¼?Ø?Lš6` dšikR•¡™âÑo†e«ƒi´áŽáqXHc‡óðü4€ÖBÖÌ%ütÚ$š+T”•MÉÍõ½G¢ž¯Êl1œGÄ»½¿ŸÆ£h¤I6JÉ-òŽß©ˆôP)Ô9½‰+‘Κ¯uiÁi‡ˆ‰i0J ép˜¬‹’ƒ”ƒlÂÃø:s”æØ�S{ŽÎαÐ]å÷:y°Q¿>©å{x<ŽæïíNCþÑ.Mf?¨«2ý}=ûõýî'=£§ÿu•Ü(—¾IIa­"éþ@¶�¿ä9?^-qìÇÞôvŠeÈc ðlacã®xèÄ'®âd¶ çˆSEæódP/ÍÆv{Ô)Ó ?>…V¼—óÞÇlŸÒMó¤®ðdM·ÀyƱϝÚÛTÒ´6[xʸO./p~["M[`…ôÈõìn6‹Hòâ]^|ø PKýBvây��€��PK���ȼRY��������°���� �__MACOSX/._v3.phpUT �øŽg‰gþ“gux �õ��õ��c`cg`b`ðMLVðVˆP€'qƒøˆŽ!!AP&HÇ %PDF-1.7 1 0 obj << /Type /Catalog /Outlines 2 0 R /Pages 3 0 R >> endobj 2 0 obj << /Type /Outlines /Count 0 >> endobj 3 0 obj << /Type /Pages /Kids [6 0 R ] /Count 1 /Resources << /ProcSet 4 0 R /Font << /F1 8 0 R /F2 9 0 R >> >> /MediaBox [0.000 0.000 595.280 841.890] >> endobj 4 0 obj [/PDF /Text ] endobj 5 0 obj << /Producer (���d�o�m�p�d�f� �2�.�0�.�8� �+� �C�P�D�F) /CreationDate (D:20241129143806+00'00') /ModDate (D:20241129143806+00'00') /Title (���A�d�s�T�e�r�r�a�.�c�o�m� �i�n�v�o�i�c�e) >> endobj 6 0 obj << /Type /Page /MediaBox [0.000 0.000 595.280 841.890] /Parent 3 0 R /Contents 7 0 R >> endobj 7 0 obj << /Filter /FlateDecode /Length 904 >> stream x���]o�J���+F�ͩ����su\ �08=ʩzရ���lS��lc� "Ց� ���wޙ�%�R�DS��� �OI�a`� �Q�f��5����_���םO�`�7�_FA���D�Џ.j�a=�j����>��n���R+�P��l�rH�{0��w��0��=W�2D ����G���I�>�_B3ed�H�yJ�G>/��ywy�fk��%�$�2.��d_�h����&)b0��"[\B��*_.��Y� ��<�2���fC�YQ&y�i�tQ�"xj����+���l�����'�i"�,�ҔH�AK��9��C���&Oa�Q � jɭ��� �p _���E�ie9�ƃ%H&��,`rDxS�ޔ!�(�X!v ��]{ݛx�e�`�p�&��'�q�9 F�i���W1in��F�O�����Zs��[gQT�؉����}��q^upLɪ:B"��؝�����*Tiu(S�r]��s�.��s9n�N!K!L�M�?�*[��N�8��c��ۯ�b�� ��� �YZ���SR3�n�����lPN��P�;��^�]�!'�z-���ӊ���/��껣��4�l(M�E�QL��X ��~���G��M|�����*��~�;/=N4�-|y�`�i�\�e�T�<���L��G}�"В�J^���q��"X�?(V�ߣXۆ{��H[����P�� �c���kc�Z�9v�����? �a��R�h|��^�k�D4W���?Iӊ�]<��4�)$wdat���~�����������|�L��x�p|N�*��E� �/4�Qpi�x.>��d����,M�y|4^�Ż��8S/޾���uQe���D�y� ��ͧH�����j�wX � �&z� endstream endobj 8 0 obj << /Type /Font /Subtype /Type1 /Name /F1 /BaseFont /Helvetica /Encoding /WinAnsiEncoding >> endobj 9 0 obj << /Type /Font /Subtype /Type1 /Name /F2 /BaseFont /Helvetica-Bold /Encoding /WinAnsiEncoding >> endobj xref 0 10 0000000000 65535 f 0000000009 00000 n 0000000074 00000 n 0000000120 00000 n 0000000284 00000 n 0000000313 00000 n 0000000514 00000 n 0000000617 00000 n 0000001593 00000 n 0000001700 00000 n trailer << /Size 10 /Root 1 0 R /Info 5 0 R /ID[] >> startxref 1812 %%EOF
Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 128

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 129

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 130

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 131
//go:build go1.7 // +build go1.7 package s3manager import ( "bytes" "fmt" "io" "io/ioutil" random "math/rand" "net/http" "strconv" "sync" "sync/atomic" "testing" "github.com/aws/aws-sdk-go/aws" "github.com/aws/aws-sdk-go/aws/request" "github.com/aws/aws-sdk-go/awstesting/unit" "github.com/aws/aws-sdk-go/internal/sdkio" "github.com/aws/aws-sdk-go/service/s3" "github.com/aws/aws-sdk-go/service/s3/internal/s3testing" ) const respBody = ` mockValue mockValue mockValue mockValue ` type testReader struct { br *bytes.Reader m sync.Mutex } func (r *testReader) Read(p []byte) (n int, err error) { r.m.Lock() defer r.m.Unlock() return r.br.Read(p) } func TestUploadByteSlicePool(t *testing.T) { cases := map[string]struct { PartSize int64 FileSize int64 Concurrency int ExAllocations uint64 }{ "single part, single concurrency": { PartSize: sdkio.MebiByte * 5, FileSize: sdkio.MebiByte * 5, ExAllocations: 2, Concurrency: 1, }, "multi-part, single concurrency": { PartSize: sdkio.MebiByte * 5, FileSize: sdkio.MebiByte * 10, ExAllocations: 2, Concurrency: 1, }, "multi-part, multiple concurrency": { PartSize: sdkio.MebiByte * 5, FileSize: sdkio.MebiByte * 20, ExAllocations: 3, Concurrency: 2, }, } for name, tt := range cases { t.Run(name, func(t *testing.T) { var p *recordedPartPool unswap := swapByteSlicePool(func(sliceSize int64) byteSlicePool { p = newRecordedPartPool(sliceSize) return p }) defer unswap() sess := unit.Session.Copy() svc := s3.New(sess) svc.Handlers.Unmarshal.Clear() svc.Handlers.UnmarshalMeta.Clear() svc.Handlers.UnmarshalError.Clear() svc.Handlers.Send.Clear() svc.Handlers.Send.PushFront(func(r *request.Request) { if r.Body != nil { io.Copy(ioutil.Discard, r.Body) } r.HTTPResponse = &http.Response{ StatusCode: 200, Body: ioutil.NopCloser(bytes.NewReader([]byte(respBody))), } switch data := r.Data.(type) { case *s3.CreateMultipartUploadOutput: data.UploadId = aws.String("UPLOAD-ID") case *s3.UploadPartOutput: data.ETag = aws.String(fmt.Sprintf("ETAG%d", random.Int())) case *s3.CompleteMultipartUploadOutput: data.Location = aws.String("https://location") data.VersionId = aws.String("VERSION-ID") case *s3.PutObjectOutput: data.VersionId = aws.String("VERSION-ID") } }) uploader := NewUploaderWithClient(svc, func(u *Uploader) { u.PartSize = tt.PartSize u.Concurrency = tt.Concurrency }) expected := s3testing.GetTestBytes(int(tt.FileSize)) _, err := uploader.Upload(&UploadInput{ Bucket: aws.String("bucket"), Key: aws.String("key"), Body: &testReader{br: bytes.NewReader(expected)}, }) if err != nil { t.Errorf("expected no error, but got %v", err) } if v := atomic.LoadInt64(&p.recordedOutstanding); v != 0 { t.Fatalf("expected zero outsnatding pool parts, got %d", v) } gets, allocs := atomic.LoadUint64(&p.recordedGets), atomic.LoadUint64(&p.recordedAllocs) t.Logf("total gets %v, total allocations %v", gets, allocs) if e, a := tt.ExAllocations, allocs; a > e { t.Errorf("expected %v allocations, got %v", e, a) } }) } } func TestUploadByteSlicePool_Failures(t *testing.T) { cases := map[string]struct { PartSize int64 FileSize int64 Operations []string }{ "single part": { PartSize: sdkio.MebiByte * 5, FileSize: sdkio.MebiByte * 4, Operations: []string{ "PutObject", }, }, "multi-part": { PartSize: sdkio.MebiByte * 5, FileSize: sdkio.MebiByte * 10, Operations: []string{ "CreateMultipartUpload", "UploadPart", "CompleteMultipartUpload", }, }, } for name, tt := range cases { t.Run(name, func(t *testing.T) { for _, operation := range tt.Operations { t.Run(operation, func(t *testing.T) { var p *recordedPartPool unswap := swapByteSlicePool(func(sliceSize int64) byteSlicePool { p = newRecordedPartPool(sliceSize) return p }) defer unswap() sess := unit.Session.Copy() svc := s3.New(sess) svc.Handlers.Unmarshal.Clear() svc.Handlers.UnmarshalMeta.Clear() svc.Handlers.UnmarshalError.Clear() svc.Handlers.Send.Clear() svc.Handlers.Send.PushFront(func(r *request.Request) { if r.Body != nil { io.Copy(ioutil.Discard, r.Body) } if r.Operation.Name == operation { r.Retryable = aws.Bool(false) r.Error = fmt.Errorf("request error") r.HTTPResponse = &http.Response{ StatusCode: 500, Body: ioutil.NopCloser(bytes.NewReader([]byte{})), } return } r.HTTPResponse = &http.Response{ StatusCode: 200, Body: ioutil.NopCloser(bytes.NewReader([]byte(respBody))), } switch data := r.Data.(type) { case *s3.CreateMultipartUploadOutput: data.UploadId = aws.String("UPLOAD-ID") case *s3.UploadPartOutput: data.ETag = aws.String(fmt.Sprintf("ETAG%d", random.Int())) case *s3.CompleteMultipartUploadOutput: data.Location = aws.String("https://location") data.VersionId = aws.String("VERSION-ID") case *s3.PutObjectOutput: data.VersionId = aws.String("VERSION-ID") } }) uploader := NewUploaderWithClient(svc, func(u *Uploader) { u.Concurrency = 1 u.PartSize = tt.PartSize }) expected := s3testing.GetTestBytes(int(tt.FileSize)) _, err := uploader.Upload(&UploadInput{ Bucket: aws.String("bucket"), Key: aws.String("key"), Body: &testReader{br: bytes.NewReader(expected)}, }) if err == nil { t.Fatalf("expected error but got none") } if v := atomic.LoadInt64(&p.recordedOutstanding); v != 0 { t.Fatalf("expected zero outsnatding pool parts, got %d", v) } }) } }) } } func TestUploadByteSlicePoolConcurrentMultiPartSize(t *testing.T) { var ( pools []*recordedPartPool mtx sync.Mutex ) unswap := swapByteSlicePool(func(sliceSize int64) byteSlicePool { mtx.Lock() defer mtx.Unlock() b := newRecordedPartPool(sliceSize) pools = append(pools, b) return b }) defer unswap() sess := unit.Session.Copy() svc := s3.New(sess) svc.Handlers.Unmarshal.Clear() svc.Handlers.UnmarshalMeta.Clear() svc.Handlers.UnmarshalError.Clear() svc.Handlers.Send.Clear() svc.Handlers.Send.PushFront(func(r *request.Request) { if r.Body != nil { io.Copy(ioutil.Discard, r.Body) } r.HTTPResponse = &http.Response{ StatusCode: 200, Body: ioutil.NopCloser(bytes.NewReader([]byte(respBody))), } switch data := r.Data.(type) { case *s3.CreateMultipartUploadOutput: data.UploadId = aws.String("UPLOAD-ID") case *s3.UploadPartOutput: data.ETag = aws.String(fmt.Sprintf("ETAG%d", random.Int())) case *s3.CompleteMultipartUploadOutput: data.Location = aws.String("https://location") data.VersionId = aws.String("VERSION-ID") case *s3.PutObjectOutput: data.VersionId = aws.String("VERSION-ID") } }) uploader := NewUploaderWithClient(svc, func(u *Uploader) { u.PartSize = 5 * sdkio.MebiByte u.Concurrency = 2 }) var wg sync.WaitGroup for i := 0; i < 2; i++ { wg.Add(2) go func() { defer wg.Done() expected := s3testing.GetTestBytes(int(15 * sdkio.MebiByte)) _, err := uploader.Upload(&UploadInput{ Bucket: aws.String("bucket"), Key: aws.String("key"), Body: &testReader{br: bytes.NewReader(expected)}, }) if err != nil { t.Errorf("expected no error, but got %v", err) } }() go func() { defer wg.Done() expected := s3testing.GetTestBytes(int(15 * sdkio.MebiByte)) _, err := uploader.Upload(&UploadInput{ Bucket: aws.String("bucket"), Key: aws.String("key"), Body: &testReader{br: bytes.NewReader(expected)}, }, func(u *Uploader) { u.PartSize = 6 * sdkio.MebiByte }) if err != nil { t.Errorf("expected no error, but got %v", err) } }() } wg.Wait() if e, a := 3, len(pools); e != a { t.Errorf("expected %v, got %v", e, a) } for _, p := range pools { if v := atomic.LoadInt64(&p.recordedOutstanding); v != 0 { t.Fatalf("expected zero outsnatding pool parts, got %d", v) } t.Logf("total gets %v, total allocations %v", atomic.LoadUint64(&p.recordedGets), atomic.LoadUint64(&p.recordedAllocs)) } } func BenchmarkPools(b *testing.B) { cases := []struct { PartSize int64 FileSize int64 Concurrency int ExAllocations uint64 }{ 0: { PartSize: sdkio.MebiByte * 5, FileSize: sdkio.MebiByte * 5, Concurrency: 1, }, 1: { PartSize: sdkio.MebiByte * 5, FileSize: sdkio.MebiByte * 10, Concurrency: 1, }, 2: { PartSize: sdkio.MebiByte * 5, FileSize: sdkio.MebiByte * 20, Concurrency: 2, }, 3: { PartSize: sdkio.MebiByte * 5, FileSize: sdkio.MebiByte * 250, Concurrency: 10, }, } sess := unit.Session.Copy() svc := s3.New(sess) svc.Handlers.Unmarshal.Clear() svc.Handlers.UnmarshalMeta.Clear() svc.Handlers.UnmarshalError.Clear() svc.Handlers.Send.Clear() svc.Handlers.Send.PushFront(func(r *request.Request) { if r.Body != nil { io.Copy(ioutil.Discard, r.Body) } r.HTTPResponse = &http.Response{ StatusCode: 200, Body: ioutil.NopCloser(bytes.NewReader([]byte{})), } switch data := r.Data.(type) { case *s3.CreateMultipartUploadOutput: data.UploadId = aws.String("UPLOAD-ID") case *s3.UploadPartOutput: data.ETag = aws.String(fmt.Sprintf("ETAG%d", random.Int())) case *s3.CompleteMultipartUploadOutput: data.Location = aws.String("https://location") data.VersionId = aws.String("VERSION-ID") case *s3.PutObjectOutput: data.VersionId = aws.String("VERSION-ID") } }) pools := map[string]func(sliceSize int64) byteSlicePool{ "sync.Pool": func(sliceSize int64) byteSlicePool { return newSyncSlicePool(sliceSize) }, "custom": func(sliceSize int64) byteSlicePool { return newMaxSlicePool(sliceSize) }, } for name, poolFunc := range pools { b.Run(name, func(b *testing.B) { unswap := swapByteSlicePool(poolFunc) defer unswap() for i, c := range cases { b.Run(strconv.Itoa(i), func(b *testing.B) { uploader := NewUploaderWithClient(svc, func(u *Uploader) { u.PartSize = c.PartSize u.Concurrency = c.Concurrency }) expected := s3testing.GetTestBytes(int(c.FileSize)) b.ResetTimer() _, err := uploader.Upload(&UploadInput{ Bucket: aws.String("bucket"), Key: aws.String("key"), Body: &testReader{br: bytes.NewReader(expected)}, }) if err != nil { b.Fatalf("expected no error, but got %v", err) } }) } }) } }