Skip to content

Commit 968f80e

Browse files
authored
Move top level package to internal (#5)
1 parent 7360856 commit 968f80e

File tree

8 files changed

+9
-10
lines changed

8 files changed

+9
-10
lines changed

cmd/scrape/main.go

+3-3
Original file line numberDiff line numberDiff line change
@@ -19,10 +19,10 @@ import (
1919
"os"
2020

2121
"github.com/efixler/envflags"
22-
"github.com/efixler/scrape"
2322
"github.com/efixler/scrape/database"
2423
"github.com/efixler/scrape/fetch"
2524
"github.com/efixler/scrape/fetch/trafilatura"
25+
"github.com/efixler/scrape/internal"
2626
"github.com/efixler/scrape/internal/cmd"
2727
"github.com/efixler/scrape/internal/headless"
2828
"github.com/efixler/scrape/internal/storage"
@@ -171,7 +171,7 @@ func openDatabase(dbh *database.DBHandle, ctx context.Context) {
171171
}
172172
}
173173

174-
func initFetcher(dbh *database.DBHandle) (*scrape.StorageBackedFetcher, error) {
174+
func initFetcher(dbh *database.DBHandle) (*internal.StorageBackedFetcher, error) {
175175
var err error
176176
var client fetch.Client
177177
if headlessEnabled {
@@ -185,7 +185,7 @@ func initFetcher(dbh *database.DBHandle) (*scrape.StorageBackedFetcher, error) {
185185
fetch.WithUserAgent(userAgent.Get().String()),
186186
)
187187
}
188-
fetcher, err := scrape.NewStorageBackedFetcher(
188+
fetcher, err := internal.NewStorageBackedFetcher(
189189
trafilatura.MustNew(client),
190190
storage.NewURLDataStore(dbh),
191191
)

benchmark_test.go renamed to internal/benchmark_test.go

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
//go:build benchmark
22

3-
package scrape
3+
package internal
44

55
// This file is associated with the tag "benchmark" to keep it from running
66
// during normal tests. It can take a while, and some of the tests call out,

scrape.go renamed to internal/scrape.go

+1-1
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ Package scrape provides a simple interface for fetching and storing web pages'
33
metadata and text content. The `scrape` and `scrape-server` commands provide
44
a command-line interface and a REST API, respectively.
55
*/
6-
package scrape
6+
package internal
77

88
import (
99
"context"

scrape_test.go renamed to internal/scrape_test.go

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
package scrape
1+
package internal
22

33
import (
44
"context"

internal/server/routes.go

+3-3
Original file line numberDiff line numberDiff line change
@@ -16,10 +16,10 @@ import (
1616

1717
"github.com/efixler/webutil/jsonarray"
1818

19-
"github.com/efixler/scrape"
2019
"github.com/efixler/scrape/database"
2120
"github.com/efixler/scrape/fetch"
2221
"github.com/efixler/scrape/fetch/feed"
22+
"github.com/efixler/scrape/internal"
2323
"github.com/efixler/scrape/internal/auth"
2424
"github.com/efixler/scrape/internal/server/healthchecks"
2525
"github.com/efixler/scrape/internal/storage"
@@ -67,7 +67,7 @@ func NewScrapeServer(
6767
directFetcher fetch.URLFetcher,
6868
headlessFetcher fetch.URLFetcher,
6969
) (*scrapeServer, error) {
70-
urlFetcher, err := scrape.NewStorageBackedFetcher(
70+
urlFetcher, err := internal.NewStorageBackedFetcher(
7171
directFetcher,
7272
storage.NewURLDataStore(dbh),
7373
)
@@ -283,7 +283,7 @@ func (ss *scrapeServer) delete(w http.ResponseWriter, r *http.Request) {
283283
http.Error(w, "Can't process delete request, no input data", http.StatusInternalServerError)
284284
return
285285
}
286-
deleter, ok := ss.urlFetcher.(*scrape.StorageBackedFetcher)
286+
deleter, ok := ss.urlFetcher.(*internal.StorageBackedFetcher)
287287
if !ok {
288288
http.Error(w, "Can't delete in the current configuration", http.StatusNotImplemented)
289289
return

internal/server/templates/index.html

-1
Original file line numberDiff line numberDiff line change
@@ -90,7 +90,6 @@
9090
name="data-iframe"
9191
class="data-iframe"
9292
>
93-
9493
</div>
9594
</body>
9695
</html>

0 commit comments

Comments
 (0)