package main import ( "encoding/json" "github.com/gorilla/feeds" "github.com/gorilla/mux" "github.com/jinzhu/gorm" _ "github.com/jinzhu/gorm/dialects/sqlite" "html" "html/template" "log" "net/http" "strconv" "strings" "time" ) var db *gorm.DB var templates = template.Must(template.ParseFiles("index.html", "FAQ.html")) var debugLogger *log.Logger var trackers = "&tr=udp://zer0day.to:1337/announce&tr=udp://tracker.leechers-paradise.org:6969&tr=udp://explodie.org:6969&tr=udp://tracker.opentrackr.org:1337&tr=udp://tracker.coppersurfer.tk:6969" func getDBHandle() *gorm.DB { dbInit, err := gorm.Open("sqlite3", "./nyaa.db") // Migrate the schema of Torrents dbInit.AutoMigrate(&Torrents{}, &Categories{}, &Sub_Categories{}, &Statuses{}) checkErr(err) return dbInit } func checkErr(err error) { if err != nil { debugLogger.Println(" " + err.Error()) } } func apiHandler(w http.ResponseWriter, r *http.Request) { vars := mux.Vars(r) page := vars["page"] pagenum, _ := strconv.Atoi(html.EscapeString(page)) b := CategoryJson{Torrents: []TorrentsJson{}} maxPerPage := 50 nbTorrents := 0 torrents := getAllTorrents(maxPerPage, maxPerPage*(pagenum-1)) for i, _ := range torrents { nbTorrents++ res := torrents[i].toJson() b.Torrents = append(b.Torrents, res) } b.QueryRecordCount = maxPerPage b.TotalRecordCount = nbTorrents w.Header().Set("Content-Type", "application/json") err := json.NewEncoder(w).Encode(b) if err != nil { http.Error(w, err.Error(), http.StatusInternalServerError) return } } func singleapiHandler(w http.ResponseWriter, r *http.Request) { vars := mux.Vars(r) id := vars["id"] b := CategoryJson{Torrents: []TorrentsJson{}} torrent, err := getTorrentById(id) res := torrent.toJson() b.Torrents = append(b.Torrents, res) b.QueryRecordCount = 1 b.TotalRecordCount = 1 w.Header().Set("Content-Type", "application/json") err = json.NewEncoder(w).Encode(b) if err != nil { http.Error(w, err.Error(), http.StatusInternalServerError) return } } func searchHandler(w http.ResponseWriter, r *http.Request) { vars := mux.Vars(r) page := vars["page"] // db params url maxPerPage, errConv := strconv.Atoi(r.URL.Query().Get("max")) if errConv != nil { maxPerPage = 50 // default Value maxPerPage } pagenum, _ := strconv.Atoi(html.EscapeString(page)) searchQuery := r.URL.Query().Get("q") cat := r.URL.Query().Get("c") stat := r.URL.Query().Get("s") catsSplit := strings.Split(cat, "_") // need this to prevent out of index panics var searchCatId, searchSubCatId string if len(catsSplit) == 2 { searchCatId = html.EscapeString(catsSplit[0]) searchSubCatId = html.EscapeString(catsSplit[1]) } nbTorrents := 0 b := []TorrentsJson{} torrents := getTorrents(createWhereParams("torrent_name LIKE ? AND status_id LIKE ? AND category_id LIKE ? AND sub_category_id LIKE ?", "%"+searchQuery+"%", stat+"%", searchCatId+"%", searchSubCatId+"%"), maxPerPage, maxPerPage*(pagenum-1)) for i, _ := range torrents { nbTorrents++ res := torrents[i].toJson() b = append(b, res) } htv := HomeTemplateVariables{b, getAllCategories(false), searchQuery, stat, cat, maxPerPage, nbTorrents} err := templates.ExecuteTemplate(w, "index.html", htv) if err != nil { http.Error(w, err.Error(), http.StatusInternalServerError) } } func safe(s string) template.URL { return template.URL(s) } func faqHandler(w http.ResponseWriter, r *http.Request) { err := templates.ExecuteTemplate(w, "FAQ.html", "") if err != nil { http.Error(w, err.Error(), http.StatusInternalServerError) } } func rssHandler(w http.ResponseWriter, r *http.Request) { //vars := mux.Vars(r) //category := vars["c"] // db params url //maxPerPage := 50 // default Value maxPerPage torrents := getFeeds() created := time.Now().String() if ( len(torrents) > 0 ) { created = torrents[0].Timestamp } created_as_time, err := time.Parse("2006-01-02 15:04:05", created) if err == nil { ; } feed := &feeds.Feed{ Title: "Nyaa Pantsu", Link: &feeds.Link{Href: "https://nyaa.pantsu.cat/"}, Created: created_as_time, } feed.Items = []*feeds.Item{} feed.Items = make( []*feeds.Item, len(torrents)) for i, _ := range torrents { timestamp_as_time, err := time.Parse("2006-01-02 15:04:05", torrents[i].Timestamp) if err == nil { feed.Items[i] = &feeds.Item{ // need a torrent view first //Id: URL + torrents[i].Hash, Title: torrents[i].Name, Link: &feeds.Link{Href: string(torrents[i].Magnet)}, Description: "", Created: timestamp_as_time, Updated: timestamp_as_time, } } } rss, err := feed.ToRss() if err == nil { w.Write( []byte( rss ) ) } else { http.Error(w, err.Error(), http.StatusInternalServerError) } } func rootHandler(w http.ResponseWriter, r *http.Request) { vars := mux.Vars(r) page := vars["page"] // db params url maxPerPage, errConv := strconv.Atoi(r.URL.Query().Get("max")) if errConv != nil { maxPerPage = 50 // default Value maxPerPage } nbTorrents := 0 pagenum, _ := strconv.Atoi(html.EscapeString(page)) b := []TorrentsJson{} torrents := getAllTorrents(maxPerPage, maxPerPage*(pagenum-1)) for i, _ := range torrents { nbTorrents++ res := torrents[i].toJson() b = append(b, res) } htv := HomeTemplateVariables{b, getAllCategories(false), "", "", "_", maxPerPage, nbTorrents} err := templates.ExecuteTemplate(w, "index.html", htv) if err != nil { http.Error(w, err.Error(), http.StatusInternalServerError) } } func main() { db = getDBHandle() router := mux.NewRouter() cssHandler := http.FileServer(http.Dir("./css/")) jsHandler := http.FileServer(http.Dir("./js/")) http.Handle("/css/", http.StripPrefix("/css/", cssHandler)) http.Handle("/js/", http.StripPrefix("/js/", jsHandler)) // Routes, router.HandleFunc("/", rootHandler) router.HandleFunc("/page/{page}", rootHandler) router.HandleFunc("/search", searchHandler) router.HandleFunc("/search/{page}", searchHandler) router.HandleFunc("/api/{page}", apiHandler).Methods("GET") router.HandleFunc("/api/torrent/{id}", singleapiHandler).Methods("GET") router.HandleFunc("/faq", faqHandler) router.HandleFunc("/feed.xml", rssHandler) http.Handle("/", router) // Set up server, srv := &http.Server{ Addr: "localhost:9999", WriteTimeout: 15 * time.Second, ReadTimeout: 15 * time.Second, } err := srv.ListenAndServe() checkErr(err) }