/* * Copyright (c) 2015 Alex Yatskov * Author: Alex Yatskov * * Permission is hereby granted, free of charge, to any person obtaining a copy of * this software and associated documentation files (the "Software"), to deal in * the Software without restriction, including without limitation the rights to * use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of * the Software, and to permit persons to whom the Software is furnished to do so, * subject to the following conditions: * * The above copyright notice and this permission notice shall be included in all * copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS * FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR * COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER * IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */ package main import ( "encoding/json" "io/ioutil" "log" "net/url" "strconv" "strings" "sync" "github.com/PuerkitoBio/goquery" ) type tabelogParams struct { Page int } type tabelogReview struct { Name string Address string Url string Dishes float64 Service float64 Atmosphere float64 Cost float64 Drinks float64 Latitude float64 Longitude float64 } func makeAbsUrl(base, ref string) (string, error) { b, err := url.Parse(base) if err != nil { return "", err } r, err := url.Parse(ref) if err != nil { return "", err } return b.ResolveReference(r).String(), nil } func dumpReviews(filename string, in chan tabelogReview) error { var reviews []tabelogReview for { if review, ok := <-in; ok { reviews = append(reviews, review) } else { break } } js, err := json.MarshalIndent(reviews, "", " ") if err != nil { return err } if err := ioutil.WriteFile(filename, js, 0644); err != nil { return err } return nil } func decodeReviews(in chan tabelogReview, out chan tabelogReview, gc *geoCache) { for { if review, ok := <-in; ok { coord, err := gc.decode(review.Address) if err == nil { review.Latitude = coord.Latitude review.Longitude = coord.Longitude out <- review } else { log.Printf("failed to decode address for %s (%v)", review.Url, err) } } else { close(out) return } } } func scrapeReview(url string, out chan tabelogReview, wc *webCache, wg *sync.WaitGroup) { defer wg.Done() doc, err := wc.load(url) if err != nil { log.Printf("failed to scrape review at %s (%v)", url, err) return } addresses := doc.Find("p.rd-detail-info__rst-address") if addresses.Length() != 2 { return } var review tabelogReview review.Url = url review.Name = doc.Find("a.rd-header__rst-name-main").Text() review.Address = strings.TrimSpace(addresses.First().Text()) if review.Dishes, err = strconv.ParseFloat(doc.Find("#js-rating-detail > dd:nth-child(2)").Text(), 8); err != nil { return } if review.Service, err = strconv.ParseFloat(doc.Find("#js-rating-detail > dd:nth-child(4)").Text(), 8); err != nil { return } if review.Atmosphere, err = strconv.ParseFloat(doc.Find("#js-rating-detail > dd:nth-child(6)").Text(), 8); err != nil { return } if review.Cost, err = strconv.ParseFloat(doc.Find("#js-rating-detail > dd:nth-child(8)").Text(), 8); err != nil { return } if review.Drinks, err = strconv.ParseFloat(doc.Find("#js-rating-detail > dd:nth-child(10)").Text(), 8); err != nil { return } out <- review } func scrapeIndex(url string, out chan tabelogReview, wc *webCache, wg *sync.WaitGroup) { doc, err := wc.load(url) if err != nil { log.Printf("failed to scrape index at %s (%v)", url, err) return } doc.Find("div.list-rst__header > p > a").Each(func(index int, sel *goquery.Selection) { if href, ok := sel.Attr("href"); ok { wg.Add(1) absUrl, err := makeAbsUrl(url, href) if err != nil { log.Fatal(err) } go scrapeReview(absUrl, out, wc, wg) } }) if href, ok := doc.Find("a.c-pagination__target--next").Attr("href"); ok { absUrl, err := makeAbsUrl(url, href) if err != nil { log.Fatal(err) } scrapeIndex(absUrl, out, wc, wg) } } func scrapeReviews(url string, out chan tabelogReview, wc *webCache) error { var wg sync.WaitGroup scrapeIndex(url, out, wc, &wg) wg.Wait() close(out) return nil } func scrapeTabelog(url, resultFile, webCacheDir, geoCacheFile string) error { wc, err := newWebCache(webCacheDir) if err != nil { return err } gc, err := newGeoCache(geoCacheFile) if err != nil { return err } scrapeChan := make(chan tabelogReview, 2000) decodeChan := make(chan tabelogReview, 2000) go decodeReviews(scrapeChan, decodeChan, gc) scrapeReviews(url, scrapeChan, wc) dumpReviews(resultFile, decodeChan) return gc.save() }