Compare commits
2 commits
3ad4c37d97
...
84d790a56e
Author | SHA1 | Date | |
---|---|---|---|
|
84d790a56e | ||
|
969dfe3fe3 |
2 changed files with 103 additions and 57 deletions
71
main.go
71
main.go
|
@ -77,38 +77,7 @@ func main() {
|
|||
}
|
||||
|
||||
func handleSearch(w http.ResponseWriter, r *http.Request) {
|
||||
var query, safe, lang, searchType string
|
||||
var page int
|
||||
|
||||
if r.Method == "GET" {
|
||||
query = r.URL.Query().Get("q")
|
||||
safe = r.URL.Query().Get("safe")
|
||||
lang = r.URL.Query().Get("lang")
|
||||
searchType = r.URL.Query().Get("t")
|
||||
pageStr := r.URL.Query().Get("p")
|
||||
var err error
|
||||
page, err = strconv.Atoi(pageStr)
|
||||
if err != nil || page < 1 {
|
||||
if debugMode {
|
||||
log.Printf("Invalid page parameter: %v, defaulting to page 1", err)
|
||||
}
|
||||
page = 1 // Default to page 1 if no valid page is specified
|
||||
}
|
||||
} else if r.Method == "POST" {
|
||||
query = r.FormValue("q")
|
||||
safe = r.FormValue("safe")
|
||||
lang = r.FormValue("lang")
|
||||
searchType = r.FormValue("t")
|
||||
pageStr := r.FormValue("p")
|
||||
var err error
|
||||
page, err = strconv.Atoi(pageStr)
|
||||
if err != nil || page < 1 {
|
||||
if debugMode {
|
||||
log.Printf("Invalid page parameter: %v, defaulting to page 1", err)
|
||||
}
|
||||
page = 1 // Default to page 1 if no valid page is specified
|
||||
}
|
||||
}
|
||||
query, safe, lang, searchType, page := parseSearchParams(r)
|
||||
|
||||
if query == "" {
|
||||
http.ServeFile(w, r, "templates/search.html")
|
||||
|
@ -116,15 +85,47 @@ func handleSearch(w http.ResponseWriter, r *http.Request) {
|
|||
}
|
||||
|
||||
switch searchType {
|
||||
case "text":
|
||||
HandleTextSearch(w, query, safe, lang, page)
|
||||
case "image":
|
||||
handleImageSearch(w, query, safe, lang, page)
|
||||
case "video":
|
||||
videoSearchEndpointHandler(w, r)
|
||||
case "map":
|
||||
handleMapSearch(w, query, safe)
|
||||
case "text":
|
||||
fallthrough
|
||||
default:
|
||||
http.ServeFile(w, r, "templates/search.html")
|
||||
HandleTextSearch(w, query, safe, lang, page)
|
||||
}
|
||||
}
|
||||
|
||||
func parseSearchParams(r *http.Request) (query, safe, lang, searchType string, page int) {
|
||||
if r.Method == "GET" {
|
||||
query = r.URL.Query().Get("q")
|
||||
safe = r.URL.Query().Get("safe")
|
||||
lang = r.URL.Query().Get("lang")
|
||||
searchType = r.URL.Query().Get("t")
|
||||
pageStr := r.URL.Query().Get("p")
|
||||
page = parsePageParameter(pageStr)
|
||||
} else if r.Method == "POST" {
|
||||
query = r.FormValue("q")
|
||||
safe = r.FormValue("safe")
|
||||
lang = r.FormValue("lang")
|
||||
searchType = r.FormValue("t")
|
||||
pageStr := r.FormValue("p")
|
||||
page = parsePageParameter(pageStr)
|
||||
}
|
||||
|
||||
if searchType == "" {
|
||||
searchType = "text"
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
func parsePageParameter(pageStr string) int {
|
||||
page, err := strconv.Atoi(pageStr)
|
||||
if err != nil || page < 1 {
|
||||
page = 1
|
||||
}
|
||||
return page
|
||||
}
|
||||
|
|
89
text.go
89
text.go
|
@ -26,9 +26,20 @@ func HandleTextSearch(w http.ResponseWriter, query, safe, lang string, page int)
|
|||
const resultsPerPage = 10
|
||||
|
||||
cacheKey := CacheKey{Query: query, Page: page, Safe: safe, Lang: lang}
|
||||
combinedResults := getResultsFromCacheOrFetch(cacheKey, query, safe, lang, page, resultsPerPage)
|
||||
|
||||
hasPrevPage := page > 1
|
||||
hasNextPage := len(combinedResults) == resultsPerPage
|
||||
|
||||
displayResults(w, combinedResults, query, lang, time.Since(startTime).Seconds(), page, hasPrevPage, hasNextPage)
|
||||
|
||||
// Always check and cache the next page
|
||||
go cacheNextPageIfNotCached(query, safe, lang, page+1, resultsPerPage)
|
||||
}
|
||||
|
||||
func getResultsFromCacheOrFetch(cacheKey CacheKey, query, safe, lang string, page, resultsPerPage int) []TextSearchResult {
|
||||
cacheChan := make(chan []TextSearchResult)
|
||||
var combinedResults []TextSearchResult
|
||||
var fromCache bool
|
||||
|
||||
go func() {
|
||||
results, exists := resultsCache.Get(cacheKey)
|
||||
|
@ -43,30 +54,55 @@ func HandleTextSearch(w http.ResponseWriter, query, safe, lang string, page int)
|
|||
|
||||
select {
|
||||
case combinedResults = <-cacheChan:
|
||||
if combinedResults != nil {
|
||||
fromCache = true
|
||||
} else {
|
||||
combinedResults = fetchAndCacheResults(query, safe, lang, page, resultsPerPage)
|
||||
if combinedResults == nil {
|
||||
combinedResults = fetchResultsUntilFull(query, safe, lang, page, resultsPerPage)
|
||||
resultsCache.Set(cacheKey, combinedResults)
|
||||
}
|
||||
case <-time.After(2 * time.Second):
|
||||
log.Println("Cache check timeout")
|
||||
combinedResults = fetchAndCacheResults(query, safe, lang, page, resultsPerPage)
|
||||
combinedResults = fetchResultsUntilFull(query, safe, lang, page, resultsPerPage)
|
||||
resultsCache.Set(cacheKey, combinedResults)
|
||||
}
|
||||
|
||||
// Only pre-fetch and cache results for the next page if we fetched new results
|
||||
if !fromCache {
|
||||
go func() {
|
||||
nextPageResults := fetchAndCacheResults(query, safe, lang, page+1, resultsPerPage)
|
||||
resultsCache.Set(CacheKey{Query: query, Page: page + 1, Safe: safe, Lang: lang}, nextPageResults)
|
||||
}()
|
||||
return combinedResults
|
||||
}
|
||||
|
||||
hasPrevPage := page > 1
|
||||
hasNextPage := len(combinedResults) == resultsPerPage
|
||||
func cacheNextPageIfNotCached(query, safe, lang string, page, resultsPerPage int) {
|
||||
cacheKey := CacheKey{Query: query, Page: page, Safe: safe, Lang: lang}
|
||||
if _, exists := resultsCache.Get(cacheKey); !exists {
|
||||
log.Printf("Next page %d not cached, caching now...", page)
|
||||
nextPageResults := fetchResultsUntilFull(query, safe, lang, page, resultsPerPage)
|
||||
resultsCache.Set(cacheKey, nextPageResults)
|
||||
} else {
|
||||
log.Printf("Next page %d already cached", page)
|
||||
}
|
||||
}
|
||||
|
||||
displayResults(w, combinedResults, query, lang, time.Since(startTime).Seconds(), page, hasPrevPage, hasNextPage)
|
||||
func fetchResultsUntilFull(query, safe, lang string, page, resultsPerPage int) []TextSearchResult {
|
||||
var combinedResults []TextSearchResult
|
||||
currentPage := 1
|
||||
resultsNeeded := page * resultsPerPage
|
||||
|
||||
for len(combinedResults) < resultsNeeded {
|
||||
results := fetchAndCacheResults(query, safe, lang, currentPage, resultsPerPage)
|
||||
if len(results) == 0 {
|
||||
break
|
||||
}
|
||||
combinedResults = append(combinedResults, results...)
|
||||
currentPage++
|
||||
}
|
||||
|
||||
startIndex := (page - 1) * resultsPerPage
|
||||
endIndex := startIndex + resultsPerPage
|
||||
|
||||
if startIndex >= len(combinedResults) {
|
||||
return []TextSearchResult{}
|
||||
}
|
||||
if endIndex > len(combinedResults) {
|
||||
endIndex = len(combinedResults)
|
||||
}
|
||||
|
||||
return combinedResults[startIndex:endIndex]
|
||||
}
|
||||
|
||||
func fetchAndCacheResults(query, safe, lang string, page, resultsPerPage int) []TextSearchResult {
|
||||
|
@ -113,24 +149,29 @@ func fetchAndCacheResults(query, safe, lang string, page, resultsPerPage int) []
|
|||
mu.Unlock()
|
||||
}
|
||||
|
||||
// Sort combinedResults by source priority: Google first, DuckDuckGo second, Qwant third
|
||||
sort.SliceStable(combinedResults, func(i, j int) bool {
|
||||
return sourceOrder(combinedResults[i].Source) < sourceOrder(combinedResults[j].Source)
|
||||
})
|
||||
|
||||
// Paginate results
|
||||
log.Printf("Fetched %d results for page %d", len(combinedResults), page)
|
||||
|
||||
return combinedResults
|
||||
}
|
||||
|
||||
func paginateResults(results []TextSearchResult, page, resultsPerPage int) []TextSearchResult {
|
||||
startIndex := (page - 1) * resultsPerPage
|
||||
endIndex := startIndex + resultsPerPage
|
||||
|
||||
// Ensure startIndex and endIndex are within bounds
|
||||
if startIndex >= len(combinedResults) {
|
||||
log.Printf("Paginating results: startIndex=%d, endIndex=%d, totalResults=%d", startIndex, endIndex, len(results))
|
||||
|
||||
if startIndex >= len(results) {
|
||||
return []TextSearchResult{}
|
||||
}
|
||||
if endIndex > len(combinedResults) {
|
||||
endIndex = len(combinedResults)
|
||||
if endIndex > len(results) {
|
||||
endIndex = len(results)
|
||||
}
|
||||
|
||||
return combinedResults[startIndex:endIndex]
|
||||
return results[startIndex:endIndex]
|
||||
}
|
||||
|
||||
func sourceOrder(source string) int {
|
||||
|
@ -147,6 +188,10 @@ func sourceOrder(source string) int {
|
|||
}
|
||||
|
||||
func displayResults(w http.ResponseWriter, results []TextSearchResult, query, lang string, elapsed float64, page int, hasPrevPage, hasNextPage bool) {
|
||||
log.Printf("Displaying results for page %d", page)
|
||||
log.Printf("Total results: %d", len(results))
|
||||
log.Printf("Has previous page: %t, Has next page: %t", hasPrevPage, hasNextPage)
|
||||
|
||||
tmpl, err := template.New("text.html").Funcs(template.FuncMap{
|
||||
"sub": func(a, b int) int {
|
||||
return a - b
|
||||
|
|
Loading…
Reference in a new issue