@@ -19,6 +19,7 @@ import (
19
19
"context"
20
20
"crypto/tls"
21
21
"crypto/x509"
22
+ "errors"
22
23
"fmt"
23
24
"io"
24
25
"net"
@@ -33,7 +34,6 @@ import (
33
34
"github.com/cenkalti/backoff/v4"
34
35
"github.com/go-kit/log"
35
36
"github.com/go-kit/log/level"
36
- "github.com/pkg/errors"
37
37
"github.com/prometheus-community/pushprox/util"
38
38
"github.com/prometheus/client_golang/prometheus"
39
39
"github.com/prometheus/client_golang/prometheus/promhttp"
@@ -134,8 +134,7 @@ func (c *Coordinator) doScrape(request *http.Request, client *http.Client) {
134
134
135
135
scrapeResp , err := client .Do (request )
136
136
if err != nil {
137
- msg := fmt .Sprintf ("failed to scrape %s" , request .URL .String ())
138
- c .handleErr (request , client , errors .Wrap (err , msg ))
137
+ c .handleErr (request , client , fmt .Errorf ("failed to scrape %s: %w" , request .URL .String (), err ))
139
138
return
140
139
}
141
140
level .Info (logger ).Log ("msg" , "Retrieved scrape response" )
@@ -184,25 +183,25 @@ func (c *Coordinator) doPoll(client *http.Client) error {
184
183
base , err := url .Parse (* proxyURL )
185
184
if err != nil {
186
185
level .Error (c .logger ).Log ("msg" , "Error parsing url:" , "err" , err )
187
- return errors . Wrap ( err , "error parsing url" )
186
+ return fmt . Errorf ( "error parsing url: %w" , err )
188
187
}
189
188
u , err := url .Parse ("poll" )
190
189
if err != nil {
191
190
level .Error (c .logger ).Log ("msg" , "Error parsing url:" , "err" , err )
192
- return errors . Wrap ( err , "error parsing url poll" )
191
+ return fmt . Errorf ( "error parsing url poll: %w" , err )
193
192
}
194
193
url := base .ResolveReference (u )
195
194
resp , err := client .Post (url .String (), "" , strings .NewReader (* myFqdn ))
196
195
if err != nil {
197
196
level .Error (c .logger ).Log ("msg" , "Error polling:" , "err" , err )
198
- return errors . Wrap ( err , "error polling" )
197
+ return fmt . Errorf ( "error polling: %w" , err )
199
198
}
200
199
defer resp .Body .Close ()
201
200
202
201
request , err := http .ReadRequest (bufio .NewReader (resp .Body ))
203
202
if err != nil {
204
203
level .Error (c .logger ).Log ("msg" , "Error reading request:" , "err" , err )
205
- return errors . Wrap ( err , "error reading request" )
204
+ return fmt . Errorf ( "error reading request: %w" , err )
206
205
}
207
206
level .Info (c .logger ).Log ("msg" , "Got scrape request" , "scrape_id" , request .Header .Get ("id" ), "url" , request .URL )
208
207
0 commit comments