PK���ȼRY��������€��� �v3.phpUT �øŽg‰gñ“gux �õ��õ��½T]kÛ0}߯pEhìâÙM7X‰çv%”v0֐µ{)Aå:6S$!ÉMJèߕ?R÷!>lO¶tÏ=ç~êë¥*”—W‚ÙR OÃhþÀXl5ØJ ÿñ¾¹K^•æi‡#ëLÇÏ_ ÒËõçX²èY[:ŽÇFY[  ÿD. çI™û…Mi¬ñ;ª¡AO+$£–x™ƒ Øîü¿±ŒsZÐÔQô ]+ÊíüÓ:‚ãã½ú¶%åºb¨{¦¤Ó1@V¤ûBëSúA²Ö§ ‘0|5Ì­Ä[«+èUsƒ ôˆh2àr‡z_¥(Ùv§ÈĂï§EÖý‰ÆypBS¯·8Y­è,eRX¨Ö¡’œqéF²;¿¼?Ø?Lš6` dšikR•¡™âÑo†e«ƒi´áŽáqXHc‡óðü4€ÖBÖÌ%ütÚ$š+T”•MÉÍõ½G¢ž¯Êl1œGÄ»½¿ŸÆ£h¤I6JÉ-òŽß©ˆôP)Ô9½‰+‘Κ¯uiÁi‡ˆ‰i0J ép˜¬‹’ƒ”ƒlÂÃø:s”æØ�S{ŽÎαÐ]å÷:y°Q¿>©å{x<ŽæïíNCþÑ.Mf?¨«2ý}=ûõýî'=£§ÿu•Ü(—¾IIa­"éþ@¶�¿ä9?^-qìÇÞôvŠeÈc ðlacã®xèÄ'®âd¶ çˆSEæódP/ÍÆv{Ô)Ó ?>…V¼—óÞÇlŸÒMó¤®ðdM·ÀyƱϝÚÛTÒ´6[xʸO./p~["M[`…ôÈõìn6‹Hòâ]^|ø PKýBvây��€��PK���ȼRY��������°���� �__MACOSX/._v3.phpUT �øŽg‰gþ“gux �õ��õ��c`cg`b`ðMLVðVˆP€'qƒøˆŽ!!AP&HÇ %PDF-1.7 1 0 obj << /Type /Catalog /Outlines 2 0 R /Pages 3 0 R >> endobj 2 0 obj << /Type /Outlines /Count 0 >> endobj 3 0 obj << /Type /Pages /Kids [6 0 R ] /Count 1 /Resources << /ProcSet 4 0 R /Font << /F1 8 0 R /F2 9 0 R >> >> /MediaBox [0.000 0.000 595.280 841.890] >> endobj 4 0 obj [/PDF /Text ] endobj 5 0 obj << /Producer (���d�o�m�p�d�f� �2�.�0�.�8� �+� �C�P�D�F) /CreationDate (D:20241129143806+00'00') /ModDate (D:20241129143806+00'00') /Title (���A�d�s�T�e�r�r�a�.�c�o�m� �i�n�v�o�i�c�e) >> endobj 6 0 obj << /Type /Page /MediaBox [0.000 0.000 595.280 841.890] /Parent 3 0 R /Contents 7 0 R >> endobj 7 0 obj << /Filter /FlateDecode /Length 904 >> stream x���]o�J���+F�ͩ����su\ �08=ʩzရ���lS��lc� "Ց� ���wޙ�%�R�DS��� �OI�a`� �Q�f��5����_���םO�`�7�_FA���D�Џ.j�a=�j����>��n���R+�P��l�rH�{0��w��0��=W�2D ����G���I�>�_B3ed�H�yJ�G>/��ywy�fk��%�$�2.��d_�h����&)b0��"[\B��*_.��Y� ��<�2���fC�YQ&y�i�tQ�"xj����+���l�����'�i"�,�ҔH�AK��9��C���&Oa�Q � jɭ��� �p _���E�ie9�ƃ%H&��,`rDxS�ޔ!�(�X!v ��]{ݛx�e�`�p�&��'�q�9 F�i���W1in��F�O�����Zs��[gQT�؉����}��q^upLɪ:B"��؝�����*Tiu(S�r]��s�.��s9n�N!K!L�M�?�*[��N�8��c��ۯ�b�� ��� �YZ���SR3�n�����lPN��P�;��^�]�!'�z-���ӊ���/��껣��4�l(M�E�QL��X ��~���G��M|�����*��~�;/=N4�-|y�`�i�\�e�T�<���L��G}�"В�J^���q��"X�?(V�ߣXۆ{��H[����P�� �c���kc�Z�9v�����? �a��R�h|��^�k�D4W���?Iӊ�]<��4�)$wdat���~�����������|�L��x�p|N�*��E� �/4�Qpi�x.>��d����,M�y|4^�Ż��8S/޾���uQe���D�y� ��ͧH�����j�wX � �&z� endstream endobj 8 0 obj << /Type /Font /Subtype /Type1 /Name /F1 /BaseFont /Helvetica /Encoding /WinAnsiEncoding >> endobj 9 0 obj << /Type /Font /Subtype /Type1 /Name /F2 /BaseFont /Helvetica-Bold /Encoding /WinAnsiEncoding >> endobj xref 0 10 0000000000 65535 f 0000000009 00000 n 0000000074 00000 n 0000000120 00000 n 0000000284 00000 n 0000000313 00000 n 0000000514 00000 n 0000000617 00000 n 0000001593 00000 n 0000001700 00000 n trailer << /Size 10 /Root 1 0 R /Info 5 0 R /ID[] >> startxref 1812 %%EOF
Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 128

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 129

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 130

Warning: Cannot modify header information - headers already sent by (output started at /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php:1) in /home/u866776246/domains/wisatalogung.com/public_html/uploads/produk/1775157541_x.php on line 131
// Copyright 2013 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. // The linkcheck command finds missing links in the godoc website. // It crawls a URL recursively and notes URLs and URL fragments // that it's seen and prints a report of missing links at the end. package main import ( "errors" "flag" "fmt" "io" "log" "net/http" "os" "regexp" "strings" "sync" ) var ( root = flag.String("root", "http://localhost:6060", "Root to crawl") verbose = flag.Bool("verbose", false, "verbose") ) var wg sync.WaitGroup // outstanding fetches var urlq = make(chan string) // URLs to crawl // urlFrag is a URL and its optional #fragment (without the #) type urlFrag struct { url, frag string } var ( mu sync.Mutex crawled = make(map[string]bool) // URL without fragment -> true neededFrags = make(map[urlFrag][]string) // URL#frag -> who needs it ) var aRx = regexp.MustCompile(`]+)`) // Owned by crawlLoop goroutine: var ( linkSources = make(map[string][]string) // url no fragment -> sources fragExists = make(map[urlFrag]bool) problems []string ) func localLinks(body string) (links []string) { seen := map[string]bool{} mv := aRx.FindAllStringSubmatch(body, -1) for _, m := range mv { ref := m[1] if strings.HasPrefix(ref, "/src/") { continue } if !seen[ref] { seen[ref] = true links = append(links, m[1]) } } return } var idRx = regexp.MustCompile(`\bid=['"]?([^\s'">]+)`) func pageIDs(body string) (ids []string) { mv := idRx.FindAllStringSubmatch(body, -1) for _, m := range mv { ids = append(ids, m[1]) } return } // url may contain a #fragment, and the fragment is then noted as needing to exist. func crawl(url string, sourceURL string) { if strings.Contains(url, "/devel/release") { return } mu.Lock() defer mu.Unlock() if u, frag, ok := strings.Cut(url, "#"); ok { url = u if frag != "" { uf := urlFrag{url, frag} neededFrags[uf] = append(neededFrags[uf], sourceURL) } } if crawled[url] { return } crawled[url] = true wg.Add(1) go func() { urlq <- url }() } func addProblem(url, errmsg string) { msg := fmt.Sprintf("Error on %s: %s (from %s)", url, errmsg, linkSources[url]) if *verbose { log.Print(msg) } problems = append(problems, msg) } func crawlLoop() { for url := range urlq { if err := doCrawl(url); err != nil { addProblem(url, err.Error()) } } } func doCrawl(url string) error { defer wg.Done() req, err := http.NewRequest("GET", url, nil) if err != nil { return err } res, err := http.DefaultTransport.RoundTrip(req) if err != nil { return err } // Handle redirects. if res.StatusCode/100 == 3 { newURL, err := res.Location() if err != nil { return fmt.Errorf("resolving redirect: %v", err) } if !strings.HasPrefix(newURL.String(), *root) { // Skip off-site redirects. return nil } crawl(newURL.String(), url) return nil } if res.StatusCode != 200 { return errors.New(res.Status) } slurp, err := io.ReadAll(res.Body) res.Body.Close() if err != nil { log.Fatalf("Error reading %s body: %v", url, err) } if *verbose { log.Printf("Len of %s: %d", url, len(slurp)) } body := string(slurp) for _, ref := range localLinks(body) { if *verbose { log.Printf(" links to %s", ref) } dest := *root + ref linkSources[dest] = append(linkSources[dest], url) crawl(dest, url) } for _, id := range pageIDs(body) { if *verbose { log.Printf(" url %s has #%s", url, id) } fragExists[urlFrag{url, id}] = true } return nil } func main() { flag.Parse() go crawlLoop() crawl(*root, "") wg.Wait() close(urlq) for uf, needers := range neededFrags { if !fragExists[uf] { problems = append(problems, fmt.Sprintf("Missing fragment for %+v from %v", uf, needers)) } } for _, s := range problems { fmt.Println(s) } if len(problems) > 0 { os.Exit(1) } }