Title: | HTTP Client |
---|---|
Description: | A simple HTTP client, with tools for making HTTP requests, and mocking HTTP requests. The package is built on R6, and takes inspiration from Ruby's 'faraday' gem (<https://rubygems.org/gems/faraday>). The package name is a play on curl, the widely used command line tool for HTTP, and this package is built on top of the R package 'curl', an interface to 'libcurl' (<https://curl.se/libcurl/>). |
Authors: | Scott Chamberlain [aut, cre] |
Maintainer: | Scott Chamberlain <[email protected]> |
License: | MIT + file LICENSE |
Version: | 1.5.0 |
Built: | 2024-10-18 06:41:44 UTC |
Source: | CRAN |
An async client to work with many URLs, but all with the same HTTP method
See HttpClient()
for information on parameters.
a list, with objects of class HttpResponse()
.
Responses are returned in the order they are passed in. We print the
first 10.
HTTP requests mostly fail in ways that you are probably familiar with, including when there's a 400 response (the URL not found), and when the server made a mistake (a 500 series HTTP status code).
But requests can fail sometimes where there is no HTTP status code, and no agreed upon way to handle it other than to just fail immediately.
When a request fails when using synchronous requests (see HttpClient) you get an error message that stops your code progression immediately saying for example:
"Could not resolve host: https://foo.com"
"Failed to connect to foo.com"
"Resolving timed out after 10 milliseconds"
However, for async requests we don't want to fail immediately because that would stop the subsequent requests from occurring. Thus, when we find that a request fails for one of the reasons above we give back a HttpResponse object just like any other response, and:
capture the error message and put it in the content
slot of the
response object (thus calls to content
and parse()
work correctly)
give back a 0
HTTP status code. we handle this specially when testing
whether the request was successful or not with e.g., the success()
method
This is an R6 class from the package R6. Find out more
about R6 at https://r6.r-lib.org/. After creating an instance of an R6
class (e.g., x <- HttpClient$new(url = "https://hb.opencpu.org")
) you can
access values and methods on the object x
.
urls
(character) one or more URLs
opts
any curl options
proxies
named list of headers
auth
an object of class auth
headers
named list of headers
print()
print method for Async objects
Async$print(x, ...)
x
self
...
ignored
new()
Create a new Async object
Async$new(urls, opts, proxies, auth, headers)
A new Async
object.
get()
execute the GET
http verb for the urls
Async$get(path = NULL, query = list(), disk = NULL, stream = NULL, ...)
path
(character) URL path, appended to the base URL
query
(list) query terms, as a named list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL
(default), memory used. See curl::curl_fetch_stream()
for help
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
\dontrun{ (cc <- Async$new(urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ))) (res <- cc$get()) }
post()
execute the POST
http verb for the urls
Async$post( path = NULL, query = list(), body = NULL, encode = "multipart", disk = NULL, stream = NULL, ... )
path
(character) URL path, appended to the base URL
query
(list) query terms, as a named list
body
body as an R list
encode
one of form, multipart, json, or raw
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL
(default), memory used. See curl::curl_fetch_stream()
for help
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
put()
execute the PUT
http verb for the urls
Async$put( path = NULL, query = list(), body = NULL, encode = "multipart", disk = NULL, stream = NULL, ... )
path
(character) URL path, appended to the base URL
query
(list) query terms, as a named list
body
body as an R list
encode
one of form, multipart, json, or raw
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL
(default), memory used. See curl::curl_fetch_stream()
for help
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
patch()
execute the PATCH
http verb for the urls
Async$patch( path = NULL, query = list(), body = NULL, encode = "multipart", disk = NULL, stream = NULL, ... )
path
(character) URL path, appended to the base URL
query
(list) query terms, as a named list
body
body as an R list
encode
one of form, multipart, json, or raw
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL
(default), memory used. See curl::curl_fetch_stream()
for help
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
delete()
execute the DELETE
http verb for the urls
Async$delete( path = NULL, query = list(), body = NULL, encode = "multipart", disk = NULL, stream = NULL, ... )
path
(character) URL path, appended to the base URL
query
(list) query terms, as a named list
body
body as an R list
encode
one of form, multipart, json, or raw
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL
(default), memory used. See curl::curl_fetch_stream()
for help
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
head()
execute the HEAD
http verb for the urls
Async$head(path = NULL, ...)
path
(character) URL path, appended to the base URL
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
retry()
execute the RETRY
http verb for the urls
. see HttpRequest$retry
method for parameters
Async$retry(...)
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
verb()
execute any supported HTTP verb
Async$verb(verb, ...)
verb
(character) a supported HTTP verb: get, post, put, patch, delete, head.
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
\dontrun{ cc <- Async$new( urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ) ) (res <- cc$verb('get')) lapply(res, function(z) z$parse("UTF-8")) }
clone()
The objects of this class are cloneable with this method.
Async$clone(deep = FALSE)
deep
Whether to make a deep clone.
Other async:
AsyncQueue
,
AsyncVaried
,
HttpRequest
## Not run: cc <- Async$new( urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ) ) cc (res <- cc$get()) res[[1]] res[[1]]$url res[[1]]$success() res[[1]]$status_http() res[[1]]$response_headers res[[1]]$method res[[1]]$content res[[1]]$parse("UTF-8") lapply(res, function(z) z$parse("UTF-8")) # curl options/headers with async urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ) cc <- Async$new(urls = urls, opts = list(verbose = TRUE), headers = list(foo = "bar") ) cc (res <- cc$get()) # using auth with async dd <- Async$new( urls = rep('https://hb.opencpu.org/basic-auth/user/passwd', 3), auth = auth(user = "foo", pwd = "passwd"), opts = list(verbose = TRUE) ) dd res <- dd$get() res vapply(res, function(z) z$status_code, double(1)) vapply(res, function(z) z$success(), logical(1)) lapply(res, function(z) z$parse("UTF-8")) # failure behavior ## e.g. when a URL doesn't exist, a timeout, etc. urls <- c("http://stuffthings.gvb", "https://foo.com", "https://hb.opencpu.org/get") conn <- Async$new(urls = urls) res <- conn$get() res[[1]]$parse("UTF-8") # a failure res[[2]]$parse("UTF-8") # a failure res[[3]]$parse("UTF-8") # a success # retry urls = c("https://hb.opencpu.org/status/404", "https://hb.opencpu.org/status/429") conn <- Async$new(urls = urls) res <- conn$retry(verb="get") ## End(Not run) ## ------------------------------------------------ ## Method `Async$get` ## ------------------------------------------------ ## Not run: (cc <- Async$new(urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ))) (res <- cc$get()) ## End(Not run) ## ------------------------------------------------ ## Method `Async$verb` ## ------------------------------------------------ ## Not run: cc <- Async$new( urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ) ) (res <- cc$verb('get')) lapply(res, function(z) z$parse("UTF-8")) ## End(Not run)
## Not run: cc <- Async$new( urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ) ) cc (res <- cc$get()) res[[1]] res[[1]]$url res[[1]]$success() res[[1]]$status_http() res[[1]]$response_headers res[[1]]$method res[[1]]$content res[[1]]$parse("UTF-8") lapply(res, function(z) z$parse("UTF-8")) # curl options/headers with async urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ) cc <- Async$new(urls = urls, opts = list(verbose = TRUE), headers = list(foo = "bar") ) cc (res <- cc$get()) # using auth with async dd <- Async$new( urls = rep('https://hb.opencpu.org/basic-auth/user/passwd', 3), auth = auth(user = "foo", pwd = "passwd"), opts = list(verbose = TRUE) ) dd res <- dd$get() res vapply(res, function(z) z$status_code, double(1)) vapply(res, function(z) z$success(), logical(1)) lapply(res, function(z) z$parse("UTF-8")) # failure behavior ## e.g. when a URL doesn't exist, a timeout, etc. urls <- c("http://stuffthings.gvb", "https://foo.com", "https://hb.opencpu.org/get") conn <- Async$new(urls = urls) res <- conn$get() res[[1]]$parse("UTF-8") # a failure res[[2]]$parse("UTF-8") # a failure res[[3]]$parse("UTF-8") # a success # retry urls = c("https://hb.opencpu.org/status/404", "https://hb.opencpu.org/status/429") conn <- Async$new(urls = urls) res <- conn$retry(verb="get") ## End(Not run) ## ------------------------------------------------ ## Method `Async$get` ## ------------------------------------------------ ## Not run: (cc <- Async$new(urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ))) (res <- cc$get()) ## End(Not run) ## ------------------------------------------------ ## Method `Async$verb` ## ------------------------------------------------ ## Not run: cc <- Async$new( urls = c( 'https://hb.opencpu.org/', 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar' ) ) (res <- cc$verb('get')) lapply(res, function(z) z$parse("UTF-8")) ## End(Not run)
An AsyncQueue client
This is an R6 class from the package R6. Find out more
about R6 at https://r6.r-lib.org/. After creating an instance of an R6
class (e.g., x <- HttpClient$new(url = "https://hb.opencpu.org")
) you can
access values and methods on the object x
.
crul::AsyncVaried
-> AsyncQueue
bucket_size
(integer) number of requests to send at once
sleep
(integer) number of seconds to sleep between each bucket
req_per_min
(integer) requests per minute
print()
print method for AsyncQueue objects
AsyncQueue$print(x, ...)
x
self
...
ignored
new()
Create a new AsyncQueue
object
AsyncQueue$new( ..., .list = list(), bucket_size = 5, sleep = NULL, req_per_min = NULL )
..., .list
Any number of objects of class HttpRequest()
,
must supply inputs to one of these parameters, but not both
bucket_size
(integer) number of requests to send at once. default: 5. See Details.
sleep
(integer) seconds to sleep between buckets. default: NULL (not set)
req_per_min
(integer) maximum number of requests per minute.
if NULL
(default), its ignored
Must set either sleep
or req_per_min
. If you set
req_per_min
we calculate a new bucket_size
when $new()
is
called
A new AsyncQueue
object
request()
Execute asynchronous requests
AsyncQueue$request()
nothing, responses stored inside object, though will print messages if you choose verbose output
responses()
List responses
AsyncQueue$responses()
a list of HttpResponse
objects, empty list before
requests made
parse()
parse content
AsyncQueue$parse(encoding = "UTF-8")
encoding
(character) the encoding to use in parsing. default:"UTF-8"
character vector, empty character vector before requests made
status_code()
Get HTTP status codes for each response
AsyncQueue$status_code()
numeric vector, empty numeric vector before requests made
status()
List HTTP status objects
AsyncQueue$status()
a list of http_code
objects, empty list before requests made
content()
Get raw content for each response
AsyncQueue$content()
raw list, empty list before requests made
times()
curl request times
AsyncQueue$times()
list of named numeric vectors, empty list before requests made
clone()
The objects of this class are cloneable with this method.
AsyncQueue$clone(deep = FALSE)
deep
Whether to make a deep clone.
Other async:
Async
,
AsyncVaried
,
HttpRequest
## Not run: # Using sleep (note this works with retry requests) reqlist <- list( HttpRequest$new(url = "https://hb.opencpu.org/get")$get(), HttpRequest$new(url = "https://hb.opencpu.org/post")$post(), HttpRequest$new(url = "https://hb.opencpu.org/put")$put(), HttpRequest$new(url = "https://hb.opencpu.org/delete")$delete(), HttpRequest$new(url = "https://hb.opencpu.org/get?g=5")$get(), HttpRequest$new( url = "https://hb.opencpu.org/post")$post(body = list(y = 9)), HttpRequest$new( url = "https://hb.opencpu.org/get")$get(query = list(hello = "world")), HttpRequest$new(url = "https://ropensci.org")$get(), HttpRequest$new(url = "https://ropensci.org/about")$get(), HttpRequest$new(url = "https://ropensci.org/packages")$get(), HttpRequest$new(url = "https://ropensci.org/community")$get(), HttpRequest$new(url = "https://ropensci.org/blog")$get(), HttpRequest$new(url = "https://ropensci.org/careers")$get(), HttpRequest$new(url = "https://hb.opencpu.org/status/404")$retry("get") ) out <- AsyncQueue$new(.list = reqlist, bucket_size = 5, sleep = 3) out out$bucket_size # bucket size out$requests() # list requests out$request() # make requests out$responses() # list responses # Using requests per minute if (interactive()) { x="https://raw.githubusercontent.com/ropensci/roregistry/gh-pages/registry.json" z <- HttpClient$new(x)$get() urls <- jsonlite::fromJSON(z$parse("UTF-8"))$packages$url repos = Filter(length, regmatches(urls, gregexpr("ropensci/[A-Za-z]+", urls))) repos = unlist(repos) auth <- list(Authorization = paste("token", Sys.getenv('GITHUB_PAT'))) reqs <- lapply(repos[1:50], function(w) { HttpRequest$new(paste0("https://api.github.com/repos/", w), headers = auth)$get() }) out <- AsyncQueue$new(.list = reqs, req_per_min = 30) out out$bucket_size out$requests() out$request() out$responses() } ## End(Not run)
## Not run: # Using sleep (note this works with retry requests) reqlist <- list( HttpRequest$new(url = "https://hb.opencpu.org/get")$get(), HttpRequest$new(url = "https://hb.opencpu.org/post")$post(), HttpRequest$new(url = "https://hb.opencpu.org/put")$put(), HttpRequest$new(url = "https://hb.opencpu.org/delete")$delete(), HttpRequest$new(url = "https://hb.opencpu.org/get?g=5")$get(), HttpRequest$new( url = "https://hb.opencpu.org/post")$post(body = list(y = 9)), HttpRequest$new( url = "https://hb.opencpu.org/get")$get(query = list(hello = "world")), HttpRequest$new(url = "https://ropensci.org")$get(), HttpRequest$new(url = "https://ropensci.org/about")$get(), HttpRequest$new(url = "https://ropensci.org/packages")$get(), HttpRequest$new(url = "https://ropensci.org/community")$get(), HttpRequest$new(url = "https://ropensci.org/blog")$get(), HttpRequest$new(url = "https://ropensci.org/careers")$get(), HttpRequest$new(url = "https://hb.opencpu.org/status/404")$retry("get") ) out <- AsyncQueue$new(.list = reqlist, bucket_size = 5, sleep = 3) out out$bucket_size # bucket size out$requests() # list requests out$request() # make requests out$responses() # list responses # Using requests per minute if (interactive()) { x="https://raw.githubusercontent.com/ropensci/roregistry/gh-pages/registry.json" z <- HttpClient$new(x)$get() urls <- jsonlite::fromJSON(z$parse("UTF-8"))$packages$url repos = Filter(length, regmatches(urls, gregexpr("ropensci/[A-Za-z]+", urls))) repos = unlist(repos) auth <- list(Authorization = paste("token", Sys.getenv('GITHUB_PAT'))) reqs <- lapply(repos[1:50], function(w) { HttpRequest$new(paste0("https://api.github.com/repos/", w), headers = auth)$get() }) out <- AsyncQueue$new(.list = reqs, req_per_min = 30) out out$bucket_size out$requests() out$request() out$responses() } ## End(Not run)
An async client to do many requests, each with different URLs, curl options, etc.
An object of class AsyncVaried
with variables and methods.
HttpResponse objects are returned in the order they are passed in.
We print the first 10.
HTTP requests mostly fail in ways that you are probably familiar with, including when there's a 400 response (the URL not found), and when the server made a mistake (a 500 series HTTP status code).
But requests can fail sometimes where there is no HTTP status code, and no agreed upon way to handle it other than to just fail immediately.
When a request fails when using synchronous requests (see HttpClient) you get an error message that stops your code progression immediately saying for example:
"Could not resolve host: https://foo.com"
"Failed to connect to foo.com"
"Resolving timed out after 10 milliseconds"
However, for async requests we don't want to fail immediately because that would stop the subsequent requests from occurring. Thus, when we find that a request fails for one of the reasons above we give back a HttpResponse object just like any other response, and:
capture the error message and put it in the content
slot of the
response object (thus calls to content
and parse()
work correctly)
give back a 0
HTTP status code. we handle this specially when testing
whether the request was successful or not with e.g., the success()
method
This is an R6 class from the package R6. Find out more
about R6 at https://r6.r-lib.org/. After creating an instance of an R6
class (e.g., x <- HttpClient$new(url = "https://hb.opencpu.org")
) you can
access values and methods on the object x
.
print()
print method for AsyncVaried objects
AsyncVaried$print(x, ...)
x
self
...
ignored
new()
Create a new AsyncVaried object
AsyncVaried$new(..., .list = list())
..., .list
Any number of objects of class HttpRequest()
,
must supply inputs to one of these parameters, but not both
A new AsyncVaried
object
request()
Execute asynchronous requests
AsyncVaried$request()
nothing, responses stored inside object, though will print messages if you choose verbose output
responses()
List responses
AsyncVaried$responses()
An S3 print method is used to summarise results. unclass
the output to see the list, or index to results, e.g., [1]
, [1:3]
a list of HttpResponse
objects, empty list before
requests made
requests()
List requests
AsyncVaried$requests()
a list of HttpRequest
objects, empty list before
requests made
parse()
parse content
AsyncVaried$parse(encoding = "UTF-8")
encoding
(character) the encoding to use in parsing. default:"UTF-8"
character vector, empty character vector before requests made
status_code()
Get HTTP status codes for each response
AsyncVaried$status_code()
numeric vector, empty numeric vector before requests made
status()
List HTTP status objects
AsyncVaried$status()
a list of http_code
objects, empty list before requests made
content()
Get raw content for each response
AsyncVaried$content()
raw list, empty list before requests made
times()
curl request times
AsyncVaried$times()
list of named numeric vectors, empty list before requests made
clone()
The objects of this class are cloneable with this method.
AsyncVaried$clone(deep = FALSE)
deep
Whether to make a deep clone.
Other async:
Async
,
AsyncQueue
,
HttpRequest
## Not run: # pass in requests via ... req1 <- HttpRequest$new( url = "https://hb.opencpu.org/get", opts = list(verbose = TRUE), headers = list(foo = "bar") )$get() req2 <- HttpRequest$new(url = "https://hb.opencpu.org/post")$post() # Create an AsyncVaried object out <- AsyncVaried$new(req1, req2) # before you make requests, the methods return empty objects out$status() out$status_code() out$content() out$times() out$parse() out$responses() # make requests out$request() # access various parts ## http status objects out$status() ## status codes out$status_code() ## content (raw data) out$content() ## times out$times() ## parsed content out$parse() ## response objects out$responses() # use $verb() method to select http verb method <- "post" req1 <- HttpRequest$new( url = "https://hb.opencpu.org/post", opts = list(verbose = TRUE), headers = list(foo = "bar") )$verb(method) req2 <- HttpRequest$new(url = "https://hb.opencpu.org/post")$verb(method) out <- AsyncVaried$new(req1, req2) out out$request() out$responses() # pass in requests in a list via .list param reqlist <- list( HttpRequest$new(url = "https://hb.opencpu.org/get")$get(), HttpRequest$new(url = "https://hb.opencpu.org/post")$post(), HttpRequest$new(url = "https://hb.opencpu.org/put")$put(), HttpRequest$new(url = "https://hb.opencpu.org/delete")$delete(), HttpRequest$new(url = "https://hb.opencpu.org/get?g=5")$get(), HttpRequest$new( url = "https://hb.opencpu.org/post")$post(body = list(y = 9)), HttpRequest$new( url = "https://hb.opencpu.org/get")$get(query = list(hello = "world")) ) out <- AsyncVaried$new(.list = reqlist) out$request() out$status() out$status_code() out$content() out$times() out$parse() # using auth with async url <- "https://hb.opencpu.org/basic-auth/user/passwd" auth <- auth(user = "user", pwd = "passwd") reqlist <- list( HttpRequest$new(url = url, auth = auth)$get(), HttpRequest$new(url = url, auth = auth)$get(query = list(a=5)), HttpRequest$new(url = url, auth = auth)$get(query = list(b=3)) ) out <- AsyncVaried$new(.list = reqlist) out$request() out$status() out$parse() # failure behavior ## e.g. when a URL doesn't exist, a timeout, etc. reqlist <- list( HttpRequest$new(url = "http://stuffthings.gvb")$get(), HttpRequest$new(url = "https://hb.opencpu.org")$head(), HttpRequest$new(url = "https://hb.opencpu.org", opts = list(timeout_ms = 10))$head() ) (tmp <- AsyncVaried$new(.list = reqlist)) tmp$request() tmp$responses() tmp$parse("UTF-8") # access intemediate redirect headers dois <- c("10.7202/1045307ar", "10.1242/jeb.088898", "10.1121/1.3383963") reqlist <- list( HttpRequest$new(url = paste0("https://doi.org/", dois[1]))$get(), HttpRequest$new(url = paste0("https://doi.org/", dois[2]))$get(), HttpRequest$new(url = paste0("https://doi.org/", dois[3]))$get() ) tmp <- AsyncVaried$new(.list = reqlist) tmp$request() tmp lapply(tmp$responses(), "[[", "response_headers_all") # retry reqlist <- list( HttpRequest$new(url = "https://hb.opencpu.org/get")$get(), HttpRequest$new(url = "https://hb.opencpu.org/post")$post(), HttpRequest$new(url = "https://hb.opencpu.org/status/404")$retry("get"), HttpRequest$new(url = "https://hb.opencpu.org/status/429")$retry("get", retry_only_on = c(403, 429), times = 2) ) tmp <- AsyncVaried$new(.list = reqlist) tmp tmp$request() tmp$responses()[[3]] ## End(Not run)
## Not run: # pass in requests via ... req1 <- HttpRequest$new( url = "https://hb.opencpu.org/get", opts = list(verbose = TRUE), headers = list(foo = "bar") )$get() req2 <- HttpRequest$new(url = "https://hb.opencpu.org/post")$post() # Create an AsyncVaried object out <- AsyncVaried$new(req1, req2) # before you make requests, the methods return empty objects out$status() out$status_code() out$content() out$times() out$parse() out$responses() # make requests out$request() # access various parts ## http status objects out$status() ## status codes out$status_code() ## content (raw data) out$content() ## times out$times() ## parsed content out$parse() ## response objects out$responses() # use $verb() method to select http verb method <- "post" req1 <- HttpRequest$new( url = "https://hb.opencpu.org/post", opts = list(verbose = TRUE), headers = list(foo = "bar") )$verb(method) req2 <- HttpRequest$new(url = "https://hb.opencpu.org/post")$verb(method) out <- AsyncVaried$new(req1, req2) out out$request() out$responses() # pass in requests in a list via .list param reqlist <- list( HttpRequest$new(url = "https://hb.opencpu.org/get")$get(), HttpRequest$new(url = "https://hb.opencpu.org/post")$post(), HttpRequest$new(url = "https://hb.opencpu.org/put")$put(), HttpRequest$new(url = "https://hb.opencpu.org/delete")$delete(), HttpRequest$new(url = "https://hb.opencpu.org/get?g=5")$get(), HttpRequest$new( url = "https://hb.opencpu.org/post")$post(body = list(y = 9)), HttpRequest$new( url = "https://hb.opencpu.org/get")$get(query = list(hello = "world")) ) out <- AsyncVaried$new(.list = reqlist) out$request() out$status() out$status_code() out$content() out$times() out$parse() # using auth with async url <- "https://hb.opencpu.org/basic-auth/user/passwd" auth <- auth(user = "user", pwd = "passwd") reqlist <- list( HttpRequest$new(url = url, auth = auth)$get(), HttpRequest$new(url = url, auth = auth)$get(query = list(a=5)), HttpRequest$new(url = url, auth = auth)$get(query = list(b=3)) ) out <- AsyncVaried$new(.list = reqlist) out$request() out$status() out$parse() # failure behavior ## e.g. when a URL doesn't exist, a timeout, etc. reqlist <- list( HttpRequest$new(url = "http://stuffthings.gvb")$get(), HttpRequest$new(url = "https://hb.opencpu.org")$head(), HttpRequest$new(url = "https://hb.opencpu.org", opts = list(timeout_ms = 10))$head() ) (tmp <- AsyncVaried$new(.list = reqlist)) tmp$request() tmp$responses() tmp$parse("UTF-8") # access intemediate redirect headers dois <- c("10.7202/1045307ar", "10.1242/jeb.088898", "10.1121/1.3383963") reqlist <- list( HttpRequest$new(url = paste0("https://doi.org/", dois[1]))$get(), HttpRequest$new(url = paste0("https://doi.org/", dois[2]))$get(), HttpRequest$new(url = paste0("https://doi.org/", dois[3]))$get() ) tmp <- AsyncVaried$new(.list = reqlist) tmp$request() tmp lapply(tmp$responses(), "[[", "response_headers_all") # retry reqlist <- list( HttpRequest$new(url = "https://hb.opencpu.org/get")$get(), HttpRequest$new(url = "https://hb.opencpu.org/post")$post(), HttpRequest$new(url = "https://hb.opencpu.org/status/404")$retry("get"), HttpRequest$new(url = "https://hb.opencpu.org/status/429")$retry("get", retry_only_on = c(403, 429), times = 2) ) tmp <- AsyncVaried$new(.list = reqlist) tmp tmp$request() tmp$responses()[[3]] ## End(Not run)
Authentication
auth(user, pwd, auth = "basic")
auth(user, pwd, auth = "basic")
user |
(character) username, required. see Details. |
pwd |
(character) password, required. see Details. |
auth |
(character) authentication type, one of basic (default), digest, digest_ie, gssnegotiate, ntlm, or any. required |
Only supporting simple auth for now, OAuth later maybe.
For user
and pwd
you are required to pass in some value.
The value can be NULL
to - which is equivalent to passing in an
empty string like ""
in httr::authenticate
. You may want to pass
in NULL
for both user
and pwd
for example if you are using
gssnegotiate
auth type. See example below.
auth(user = "foo", pwd = "bar", auth = "basic") auth(user = "foo", pwd = "bar", auth = "digest") auth(user = "foo", pwd = "bar", auth = "ntlm") auth(user = "foo", pwd = "bar", auth = "any") # gssnegotiate auth auth(NULL, NULL, "gssnegotiate") ## Not run: # with HttpClient (res <- HttpClient$new( url = "https://hb.opencpu.org/basic-auth/user/passwd", auth = auth(user = "user", pwd = "passwd") )) res$auth x <- res$get() jsonlite::fromJSON(x$parse("UTF-8")) # with HttpRequest (res <- HttpRequest$new( url = "https://hb.opencpu.org/basic-auth/user/passwd", auth = auth(user = "user", pwd = "passwd") )) res$auth ## End(Not run)
auth(user = "foo", pwd = "bar", auth = "basic") auth(user = "foo", pwd = "bar", auth = "digest") auth(user = "foo", pwd = "bar", auth = "ntlm") auth(user = "foo", pwd = "bar", auth = "any") # gssnegotiate auth auth(NULL, NULL, "gssnegotiate") ## Not run: # with HttpClient (res <- HttpClient$new( url = "https://hb.opencpu.org/basic-auth/user/passwd", auth = auth(user = "user", pwd = "passwd") )) res$auth x <- res$get() jsonlite::fromJSON(x$parse("UTF-8")) # with HttpRequest (res <- HttpRequest$new( url = "https://hb.opencpu.org/basic-auth/user/passwd", auth = auth(user = "user", pwd = "passwd") )) res$auth ## End(Not run)
The HttpResponse class holds all the responses elements for an HTTP request. This document details how to work specifically with the content-type of the response headers
The "Content-Type" header in HTTP responses gives the media type of the response. The media type is both the data format and how the data is intended to be processed by a recipient. (modified from rfc7231)
type: (only applicable for the raise_for_ct()
method): instead of
using one of the three other content type methods for html, json, or xml,
you can specify a mime type to check, any of those in mime::mimemap
charset: if you don't give a value to this parameter, we only check that the content type is what you expect; that is, the charset, if given, is ignored.
behavior: by default when you call this method, and the content type
does not match what the method expects, then we run stop()
with a
message. Instead of stopping, you can choose behavior="warning"
and we'll throw a warning instead, allowing any downstream processing
to proceed.
spec for content types: https://datatracker.ietf.org/doc/html/rfc7231#section-3.1.1.5
spec for media types: https://datatracker.ietf.org/doc/html/rfc7231#section-3.1.1.1
## Not run: (x <- HttpClient$new(url = "https://hb.opencpu.org")) (res <- x$get()) ## see the content type res$response_headers ## check that the content type is text/html res$raise_for_ct_html() ## it's def. not json # res$raise_for_ct_json() ## give custom content type res$raise_for_ct("text/html") # res$raise_for_ct("application/json") # res$raise_for_ct("foo/bar") ## check charset in addition to the media type res$raise_for_ct_html(charset = "utf-8") # res$raise_for_ct_html(charset = "utf-16") # warn instead of stop res$raise_for_ct_json(behavior = "warning") ## End(Not run)
## Not run: (x <- HttpClient$new(url = "https://hb.opencpu.org")) (res <- x$get()) ## see the content type res$response_headers ## check that the content type is text/html res$raise_for_ct_html() ## it's def. not json # res$raise_for_ct_json() ## give custom content type res$raise_for_ct("text/html") # res$raise_for_ct("application/json") # res$raise_for_ct("foo/bar") ## check charset in addition to the media type res$raise_for_ct_html(charset = "utf-8") # res$raise_for_ct_html(charset = "utf-16") # warn instead of stop res$raise_for_ct_json(behavior = "warning") ## End(Not run)
Working with cookies
## Not run: x <- HttpClient$new( url = "https://hb.opencpu.org", opts = list( cookie = "c=1;f=5", verbose = TRUE ) ) x # set cookies (res <- x$get("cookies")) jsonlite::fromJSON(res$parse("UTF-8")) (x <- HttpClient$new(url = "https://hb.opencpu.org")) res <- x$get("cookies/set", query = list(foo = 123, bar = "ftw")) jsonlite::fromJSON(res$parse("UTF-8")) curl::handle_cookies(handle = res$handle) # reuse handle res2 <- x$get("get", query = list(hello = "world")) jsonlite::fromJSON(res2$parse("UTF-8")) curl::handle_cookies(handle = res2$handle) # DOAJ x <- HttpClient$new(url = "https://doaj.org") res <- x$get("api/v1/journals/f3f2e7f23d444370ae5f5199f85bc100", verbose = TRUE) res$response_headers$`set-cookie` curl::handle_cookies(handle = res$handle) res2 <- x$get("api/v1/journals/9abfb36b06404e8a8566e1a44180bbdc", verbose = TRUE) ## reset handle x$handle_pop() ## cookies no longer sent, as handle reset res2 <- x$get("api/v1/journals/9abfb36b06404e8a8566e1a44180bbdc", verbose = TRUE) ## End(Not run)
## Not run: x <- HttpClient$new( url = "https://hb.opencpu.org", opts = list( cookie = "c=1;f=5", verbose = TRUE ) ) x # set cookies (res <- x$get("cookies")) jsonlite::fromJSON(res$parse("UTF-8")) (x <- HttpClient$new(url = "https://hb.opencpu.org")) res <- x$get("cookies/set", query = list(foo = 123, bar = "ftw")) jsonlite::fromJSON(res$parse("UTF-8")) curl::handle_cookies(handle = res$handle) # reuse handle res2 <- x$get("get", query = list(hello = "world")) jsonlite::fromJSON(res2$parse("UTF-8")) curl::handle_cookies(handle = res2$handle) # DOAJ x <- HttpClient$new(url = "https://doaj.org") res <- x$get("api/v1/journals/f3f2e7f23d444370ae5f5199f85bc100", verbose = TRUE) res$response_headers$`set-cookie` curl::handle_cookies(handle = res$handle) res2 <- x$get("api/v1/journals/9abfb36b06404e8a8566e1a44180bbdc", verbose = TRUE) ## reset handle x$handle_pop() ## cookies no longer sent, as handle reset res2 <- x$get("api/v1/journals/9abfb36b06404e8a8566e1a44180bbdc", verbose = TRUE) ## End(Not run)
Set curl options, proxy, and basic auth
set_opts(...) set_verbose() set_proxy(x) set_auth(x) set_headers(...) crul_settings(reset = FALSE)
set_opts(...) set_verbose() set_proxy(x) set_auth(x) set_headers(...) crul_settings(reset = FALSE)
... |
For |
x |
For |
reset |
(logical) reset all settings (aka, delete them).
Default: |
set_opts()
: set curl options; supports any options in
curl::curl_options()
set_verbose()
: set custom curl verbose; sets verbose=TRUE
and debugfunction
to the callback result from curl_verbose()
set_proxy()
: set proxy settings, accepts proxy()
set_auth()
: set authorization, accepts auth()
set_headers()
: set request headers, a named list
crul_settings()
: list all settigns set via these functions
the mock
option will be seen in output of crul_settings()
but is set via the function mock()
if (interactive()) { # get settings crul_settings() # curl options set_opts(timeout_ms = 1000) crul_settings() set_opts(timeout_ms = 4000) crul_settings() set_opts(verbose = TRUE) crul_settings() ## Not run: HttpClient$new('https://hb.opencpu.org')$get('get') ## End(Not run) # set_verbose - sets: `verbose=TRUE`, and `debugfunction` to # result of call to `curl_verbose()`, see `?curl_verbose` set_verbose() crul_settings() # basic authentication set_auth(auth(user = "foo", pwd = "bar", auth = "basic")) crul_settings() # proxies set_proxy(proxy("http://97.77.104.22:3128")) crul_settings() # headers crul_settings(TRUE) # reset first set_headers(foo = "bar") crul_settings() set_headers(`User-Agent` = "hello world") crul_settings() ## Not run: set_opts(verbose = TRUE) HttpClient$new('https://hb.opencpu.org')$get('get') ## End(Not run) # reset crul_settings(TRUE) crul_settings() # works with async functions ## Async set_opts(verbose = TRUE) cc <- Async$new(urls = c( 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar')) (res <- cc$get()) ## AsyncVaried set_opts(verbose = TRUE) set_headers(stuff = "things") reqlist <- list( HttpRequest$new(url = "https://hb.opencpu.org/get")$get(), HttpRequest$new(url = "https://hb.opencpu.org/post")$post()) out <- AsyncVaried$new(.list = reqlist) out$request() }
if (interactive()) { # get settings crul_settings() # curl options set_opts(timeout_ms = 1000) crul_settings() set_opts(timeout_ms = 4000) crul_settings() set_opts(verbose = TRUE) crul_settings() ## Not run: HttpClient$new('https://hb.opencpu.org')$get('get') ## End(Not run) # set_verbose - sets: `verbose=TRUE`, and `debugfunction` to # result of call to `curl_verbose()`, see `?curl_verbose` set_verbose() crul_settings() # basic authentication set_auth(auth(user = "foo", pwd = "bar", auth = "basic")) crul_settings() # proxies set_proxy(proxy("http://97.77.104.22:3128")) crul_settings() # headers crul_settings(TRUE) # reset first set_headers(foo = "bar") crul_settings() set_headers(`User-Agent` = "hello world") crul_settings() ## Not run: set_opts(verbose = TRUE) HttpClient$new('https://hb.opencpu.org')$get('get') ## End(Not run) # reset crul_settings(TRUE) crul_settings() # works with async functions ## Async set_opts(verbose = TRUE) cc <- Async$new(urls = c( 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar')) (res <- cc$get()) ## AsyncVaried set_opts(verbose = TRUE) set_headers(stuff = "things") reqlist <- list( HttpRequest$new(url = "https://hb.opencpu.org/get")$get(), HttpRequest$new(url = "https://hb.opencpu.org/post")$post()) out <- AsyncVaried$new(.list = reqlist) out$request() }
curl verbose method
curl_verbose(data_out = TRUE, data_in = FALSE, info = FALSE, ssl = FALSE)
curl_verbose(data_out = TRUE, data_in = FALSE, info = FALSE, ssl = FALSE)
data_out |
Show data sent to the server |
data_in |
Show data recieved from the server |
info |
Show informational text from curl. This is mainly useful for debugging https and auth problems, so is disabled by default |
ssl |
Show even data sent/recieved over SSL connections? |
line prefixes:
*
informative curl messages
=>
headers sent (out)
>
data sent (out)
*>
ssl data sent (out)
<=
headers received (in)
<
data received (in)
<*
ssl data received (in)
adapted from httr::verbose
With the opts
parameter you can pass in various
curl options, including user agent string, whether to get verbose
curl output or not, setting a timeout for requests, and more. See
curl::curl_options()
for all the options you can use. Note that
you need to give curl options exactly as given in
curl::curl_options()
.
## Not run: url <- "https://hb.opencpu.org" # set curl options on client initialization (res <- HttpClient$new(url = url, opts = list(verbose = TRUE))) res$opts res$get('get') # or set curl options when performing HTTP operation (res <- HttpClient$new(url = url)) res$get('get', verbose = TRUE) res$get('get', stuff = "things") # set a timeout (res <- HttpClient$new(url = url, opts = list(timeout_ms = 1))) # res$get('get') # set user agent either as a header or an option HttpClient$new(url = url, headers = list(`User-Agent` = "hello world"), opts = list(verbose = TRUE) )$get('get') HttpClient$new(url = url, opts = list(verbose = TRUE, useragent = "hello world") )$get('get') # You can also set custom debug function via the verbose # parameter when calling `$new()` res <- HttpClient$new(url, verbose=curl_verbose()) res res$get("get") res <- HttpClient$new(url, verbose=curl_verbose(data_in=TRUE)) res$get("get") res <- HttpClient$new(url, verbose=curl_verbose(info=TRUE)) res$get("get") ## End(Not run)
## Not run: url <- "https://hb.opencpu.org" # set curl options on client initialization (res <- HttpClient$new(url = url, opts = list(verbose = TRUE))) res$opts res$get('get') # or set curl options when performing HTTP operation (res <- HttpClient$new(url = url)) res$get('get', verbose = TRUE) res$get('get', stuff = "things") # set a timeout (res <- HttpClient$new(url = url, opts = list(timeout_ms = 1))) # res$get('get') # set user agent either as a header or an option HttpClient$new(url = url, headers = list(`User-Agent` = "hello world"), opts = list(verbose = TRUE) )$get('get') HttpClient$new(url = url, opts = list(verbose = TRUE, useragent = "hello world") )$get('get') # You can also set custom debug function via the verbose # parameter when calling `$new()` res <- HttpClient$new(url, verbose=curl_verbose()) res res$get("get") res <- HttpClient$new(url, verbose=curl_verbose(data_in=TRUE)) res$get("get") res <- HttpClient$new(url, verbose=curl_verbose(info=TRUE)) res$get("get") ## End(Not run)
Make a handle
handle(url, ...)
handle(url, ...)
url |
(character) A url. required. |
... |
options passed on to |
handle("https://hb.opencpu.org") # handles - pass in your own handle ## Not run: h <- handle("https://hb.opencpu.org") (res <- HttpClient$new(handle = h)) out <- res$get("get") ## End(Not run)
handle("https://hb.opencpu.org") # handles - pass in your own handle ## Not run: h <- handle("https://hb.opencpu.org") (res <- HttpClient$new(handle = h)) out <- res$get("get") ## End(Not run)
Trigger functions to run on requests and/or responses. See Details for more.
Functions passed to request
are run before the
request occurs. The meaning of triggering a function on the
request is that you can do things to the request object.
Functions passed to response
are run once the
request is done, and the response object is created.
The meaning of triggering a function on the
response is to do things on the response object.
The above for request and response applies the same
whether you make real HTTP requests or mock with
webmockr
.
Only supported on HttpClient for now
## Not run: # hooks on the request fun_req <- function(request) { cat(paste0("Requesting: ", request$url$url), sep = "\n") } (x <- HttpClient$new(url = "https://hb.opencpu.org", hooks = list(request = fun_req))) x$hooks x$hooks$request r1 <- x$get('get') captured_req <- list() fun_req2 <- function(request) { cat("Capturing Request", sep = "\n") captured_req <<- request } (x <- HttpClient$new(url = "https://hb.opencpu.org", hooks = list(request = fun_req2))) x$hooks x$hooks$request r1 <- x$get('get') captured_req # hooks on the response fun_resp <- function(response) { cat(paste0("status_code: ", response$status_code), sep = "\n") } (x <- HttpClient$new(url = "https://hb.opencpu.org", hooks = list(response = fun_resp))) x$url x$hooks r1 <- x$get('get') # both (x <- HttpClient$new(url = "https://hb.opencpu.org", hooks = list(request = fun_req, response = fun_resp))) x$get("get") ## End(Not run)
## Not run: # hooks on the request fun_req <- function(request) { cat(paste0("Requesting: ", request$url$url), sep = "\n") } (x <- HttpClient$new(url = "https://hb.opencpu.org", hooks = list(request = fun_req))) x$hooks x$hooks$request r1 <- x$get('get') captured_req <- list() fun_req2 <- function(request) { cat("Capturing Request", sep = "\n") captured_req <<- request } (x <- HttpClient$new(url = "https://hb.opencpu.org", hooks = list(request = fun_req2))) x$hooks x$hooks$request r1 <- x$get('get') captured_req # hooks on the response fun_resp <- function(response) { cat(paste0("status_code: ", response$status_code), sep = "\n") } (x <- HttpClient$new(url = "https://hb.opencpu.org", hooks = list(response = fun_resp))) x$url x$hooks r1 <- x$get('get') # both (x <- HttpClient$new(url = "https://hb.opencpu.org", hooks = list(request = fun_req, response = fun_resp))) x$get("get") ## End(Not run)
Working with HTTP headers
## Not run: (x <- HttpClient$new(url = "https://hb.opencpu.org")) # set headers (res <- HttpClient$new( url = "https://hb.opencpu.org", opts = list( verbose = TRUE ), headers = list( a = "stuff", b = "things" ) )) res$headers # reassign header value res$headers$a <- "that" # define new header res$headers$c <- "what" # request res$get('get') ## setting content-type via headers (res <- HttpClient$new( url = "https://hb.opencpu.org", opts = list( verbose = TRUE ), headers = list(`Content-Type` = "application/json") )) res$get('get') ## End(Not run)
## Not run: (x <- HttpClient$new(url = "https://hb.opencpu.org")) # set headers (res <- HttpClient$new( url = "https://hb.opencpu.org", opts = list( verbose = TRUE ), headers = list( a = "stuff", b = "things" ) )) res$headers # reassign header value res$headers$a <- "that" # define new header res$headers$c <- "what" # request res$get('get') ## setting content-type via headers (res <- HttpClient$new( url = "https://hb.opencpu.org", opts = list( verbose = TRUE ), headers = list(`Content-Type` = "application/json") )) res$get('get') ## End(Not run)
Create and execute HTTP requests
an HttpResponse object
This is an R6 class from the package R6. Find out more
about R6 at https://r6.r-lib.org/. After creating an instance of an R6
class (e.g., x <- HttpClient$new(url = "https://hb.opencpu.org")
) you can
access values and methods on the object x
.
curl handles are re-used on the level of the connection object, that is,
each HttpClient
object is separate from one another so as to better
separate connections.
If you don't pass in a curl handle to the handle
parameter,
it gets created when a HTTP verb is called. Thus, if you try to get handle
after creating a HttpClient
object only passing url
parameter, handle
will be NULL
. If you pass a curl handle to the handle
parameter, then
you can get the handle from the HttpClient
object. The response from a
http verb request does have the handle in the handle
slot.
print()
print method for HttpClient
objects
HttpClient$print(x, ...)
x
self
...
ignored
new()
Create a new HttpClient object
HttpClient$new( url, opts, proxies, auth, headers, handle, progress, hooks, verbose )
url
(character) A url. One of url
or handle
required.
opts
any curl options
proxies
a proxy()
object
auth
an auth()
object
headers
named list of headers, see http-headers
handle
a handle()
progress
only supports httr::progress()
, see progress
hooks
a named list, see hooks
verbose
a special handler for verbose curl output,
accepts a function only. default is NULL
. if used, verbose
and debugfunction
curl options are ignored if passed to opts
on $new()
and ignored if ...
passed to a http method call
urls
(character) one or more URLs
A new HttpClient
object
get()
Make a GET request
HttpClient$get(path = NULL, query = list(), disk = NULL, stream = NULL, ...)
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
post()
Make a POST request
HttpClient$post( path = NULL, query = list(), body = NULL, disk = NULL, stream = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
body
body as an R list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
encode
one of form, multipart, json, or raw
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
put()
Make a PUT request
HttpClient$put( path = NULL, query = list(), body = NULL, disk = NULL, stream = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
body
body as an R list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
encode
one of form, multipart, json, or raw
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
patch()
Make a PATCH request
HttpClient$patch( path = NULL, query = list(), body = NULL, disk = NULL, stream = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
body
body as an R list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
encode
one of form, multipart, json, or raw
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
delete()
Make a DELETE request
HttpClient$delete( path = NULL, query = list(), body = NULL, disk = NULL, stream = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
body
body as an R list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
encode
one of form, multipart, json, or raw
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
head()
Make a HEAD request
HttpClient$head(path = NULL, query = list(), ...)
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
verb()
Use an arbitrary HTTP verb supported on this class Supported verbs: "get", "post", "put", "patch", "delete", "head". Also supports retry
HttpClient$verb(verb, ...)
verb
an HTTP verb supported on this class: "get", "post", "put", "patch", "delete", "head". Also supports retry.
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
\dontrun{ (x <- HttpClient$new(url = "https://hb.opencpu.org")) x$verb('get') x$verb('GET') x$verb('GET', query = list(foo = "bar")) x$verb('retry', 'GET', path = "status/400") }
retry()
Retry a request
HttpClient$retry( verb, ..., pause_base = 1, pause_cap = 60, pause_min = 1, times = 3, terminate_on = NULL, retry_only_on = NULL, onwait = NULL )
verb
an HTTP verb supported on this class: "get", "post", "put", "patch", "delete", "head". Also supports retry.
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
pause_base, pause_cap, pause_min
basis, maximum, and minimum for
calculating wait time for retry. Wait time is calculated according to the
exponential backoff with full jitter algorithm. Specifically, wait time is
chosen randomly between pause_min
and the lesser of pause_base * 2
and
pause_cap
, with pause_base
doubling on each subsequent retry attempt.
Use pause_cap = Inf
to not terminate retrying due to cap of wait time
reached.
times
the maximum number of times to retry. Set to Inf
to
not stop retrying due to exhausting the number of attempts.
terminate_on, retry_only_on
a vector of HTTP status codes. For
terminate_on
, the status codes for which to terminate retrying, and for
retry_only_on
, the status codes for which to retry the request.
onwait
a callback function if the request will be retried and a wait time is being applied. The function will be passed two parameters, the response object from the failed request, and the wait time in seconds. Note that the time spent in the function effectively adds to the wait time, so it should be kept simple.
Retries the request given by verb
until successful
(HTTP response status < 400), or a condition for giving up is met.
Automatically recognizes Retry-After
and X-RateLimit-Reset
headers
in the response for rate-limited remote APIs.
\dontrun{ x <- HttpClient$new(url = "https://hb.opencpu.org") # retry, by default at most 3 times (res_get <- x$retry("GET", path = "status/400")) # retry, but not for 404 NOT FOUND (res_get <- x$retry("GET", path = "status/404", terminate_on = c(404))) # retry, but only for exceeding rate limit (note that e.g. Github uses 403) (res_get <- x$retry("GET", path = "status/429", retry_only_on = c(403, 429))) }
handle_pop()
reset your curl handle
HttpClient$handle_pop()
url_fetch()
get the URL that would be sent (i.e., before executing the request) the only things that change the URL are path and query parameters; body and any curl options don't change the URL
HttpClient$url_fetch(path = NULL, query = list())
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
URL (character)
x <- HttpClient$new(url = "https://hb.opencpu.org") x$url_fetch() x$url_fetch('get') x$url_fetch('post') x$url_fetch('get', query = list(foo = "bar"))
clone()
The objects of this class are cloneable with this method.
HttpClient$clone(deep = FALSE)
deep
Whether to make a deep clone.
A little quirk about crul
is that because user agent string can
be passed as either a header or a curl option (both lead to a User-Agent
header being passed in the HTTP request), we return the user agent
string in the request_headers
list of the response even if you
pass in a useragent
string as a curl option. Note that whether you pass
in as a header like User-Agent
or as a curl option like useragent
,
it is returned as request_headers$User-Agent
so at least accessing
it in the request headers is consistent.
http-headers, writing-options, cookies, hooks
## Not run: # set your own handle (h <- handle("https://hb.opencpu.org")) (x <- HttpClient$new(handle = h)) x$handle x$url (out <- x$get("get")) x$handle x$url class(out) out$handle out$request_headers out$response_headers out$response_headers_all # if you just pass a url, we create a handle for you # this is how most people will use HttpClient (x <- HttpClient$new(url = "https://hb.opencpu.org")) x$url x$handle # is empty, it gets created when a HTTP verb is called (r1 <- x$get('get')) x$url x$handle r1$url r1$handle r1$content r1$response_headers r1$parse() (res_get2 <- x$get('get', query = list(hello = "world"))) res_get2$parse() library("jsonlite") jsonlite::fromJSON(res_get2$parse()) # post request (res_post <- x$post('post', body = list(hello = "world"))) ## empty body request x$post('post') # put request (res_put <- x$put('put')) # delete request (res_delete <- x$delete('delete')) # patch request (res_patch <- x$patch('patch')) # head request (res_head <- x$head()) # query params are URL encoded for you, so DO NOT do it yourself ## if you url encode yourself, it gets double encoded, and that's bad (x <- HttpClient$new(url = "https://hb.opencpu.org")) res <- x$get("get", query = list(a = 'hello world')) # access intermediate headers in response_headers_all x <- HttpClient$new("https://doi.org/10.1007/978-3-642-40455-9_52-1") bb <- x$get() bb$response_headers_all ## End(Not run) ## ------------------------------------------------ ## Method `HttpClient$verb` ## ------------------------------------------------ ## Not run: (x <- HttpClient$new(url = "https://hb.opencpu.org")) x$verb('get') x$verb('GET') x$verb('GET', query = list(foo = "bar")) x$verb('retry', 'GET', path = "status/400") ## End(Not run) ## ------------------------------------------------ ## Method `HttpClient$retry` ## ------------------------------------------------ ## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") # retry, by default at most 3 times (res_get <- x$retry("GET", path = "status/400")) # retry, but not for 404 NOT FOUND (res_get <- x$retry("GET", path = "status/404", terminate_on = c(404))) # retry, but only for exceeding rate limit (note that e.g. Github uses 403) (res_get <- x$retry("GET", path = "status/429", retry_only_on = c(403, 429))) ## End(Not run) ## ------------------------------------------------ ## Method `HttpClient$url_fetch` ## ------------------------------------------------ x <- HttpClient$new(url = "https://hb.opencpu.org") x$url_fetch() x$url_fetch('get') x$url_fetch('post') x$url_fetch('get', query = list(foo = "bar"))
## Not run: # set your own handle (h <- handle("https://hb.opencpu.org")) (x <- HttpClient$new(handle = h)) x$handle x$url (out <- x$get("get")) x$handle x$url class(out) out$handle out$request_headers out$response_headers out$response_headers_all # if you just pass a url, we create a handle for you # this is how most people will use HttpClient (x <- HttpClient$new(url = "https://hb.opencpu.org")) x$url x$handle # is empty, it gets created when a HTTP verb is called (r1 <- x$get('get')) x$url x$handle r1$url r1$handle r1$content r1$response_headers r1$parse() (res_get2 <- x$get('get', query = list(hello = "world"))) res_get2$parse() library("jsonlite") jsonlite::fromJSON(res_get2$parse()) # post request (res_post <- x$post('post', body = list(hello = "world"))) ## empty body request x$post('post') # put request (res_put <- x$put('put')) # delete request (res_delete <- x$delete('delete')) # patch request (res_patch <- x$patch('patch')) # head request (res_head <- x$head()) # query params are URL encoded for you, so DO NOT do it yourself ## if you url encode yourself, it gets double encoded, and that's bad (x <- HttpClient$new(url = "https://hb.opencpu.org")) res <- x$get("get", query = list(a = 'hello world')) # access intermediate headers in response_headers_all x <- HttpClient$new("https://doi.org/10.1007/978-3-642-40455-9_52-1") bb <- x$get() bb$response_headers_all ## End(Not run) ## ------------------------------------------------ ## Method `HttpClient$verb` ## ------------------------------------------------ ## Not run: (x <- HttpClient$new(url = "https://hb.opencpu.org")) x$verb('get') x$verb('GET') x$verb('GET', query = list(foo = "bar")) x$verb('retry', 'GET', path = "status/400") ## End(Not run) ## ------------------------------------------------ ## Method `HttpClient$retry` ## ------------------------------------------------ ## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") # retry, by default at most 3 times (res_get <- x$retry("GET", path = "status/400")) # retry, but not for 404 NOT FOUND (res_get <- x$retry("GET", path = "status/404", terminate_on = c(404))) # retry, but only for exceeding rate limit (note that e.g. Github uses 403) (res_get <- x$retry("GET", path = "status/429", retry_only_on = c(403, 429))) ## End(Not run) ## ------------------------------------------------ ## Method `HttpClient$url_fetch` ## ------------------------------------------------ x <- HttpClient$new(url = "https://hb.opencpu.org") x$url_fetch() x$url_fetch('get') x$url_fetch('post') x$url_fetch('get', query = list(foo = "bar"))
Create HTTP requests
This R6 class doesn't do actual HTTP requests as does
HttpClient()
- it is for building requests to use for async HTTP
requests in AsyncVaried()
Note that you can access HTTP verbs after creating an HttpRequest
object, just as you can with HttpClient
. See examples for usage.
Also note that when you call HTTP verbs on a HttpRequest
object you
don't need to assign the new object to a variable as the new details
you've added are added to the object itself.
See HttpClient()
for information on parameters.
This is an R6 class from the package R6. Find out more
about R6 at https://r6.r-lib.org/. After creating an instance of an R6
class (e.g., x <- HttpClient$new(url = "https://hb.opencpu.org")
) you can
access values and methods on the object x
.
url
(character) a url
opts
(list) named list of curl options
proxies
a proxy()
object
auth
an auth()
object
headers
(list) named list of headers, see http-headers
handle
a handle()
progress
only supports httr::progress()
, see progress
payload
resulting payload after request
print()
print method for HttpRequest
objects
HttpRequest$print(x, ...)
x
self
...
ignored
new()
Create a new HttpRequest
object
HttpRequest$new(url, opts, proxies, auth, headers, handle, progress)
url
(character) A url. One of url
or handle
required.
opts
any curl options
proxies
a proxy()
object
auth
an auth()
object
headers
named list of headers, see http-headers
handle
a handle()
progress
only supports httr::progress()
, see progress
urls
(character) one or more URLs
A new HttpRequest
object
get()
Define a GET request
HttpRequest$get(path = NULL, query = list(), disk = NULL, stream = NULL, ...)
path
URL path, appended to the base URL
query
query terms, as a named list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
post()
Define a POST request
HttpRequest$post( path = NULL, query = list(), body = NULL, disk = NULL, stream = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list
body
body as an R list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
encode
one of form, multipart, json, or raw
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
put()
Define a PUT request
HttpRequest$put( path = NULL, query = list(), body = NULL, disk = NULL, stream = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list
body
body as an R list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
encode
one of form, multipart, json, or raw
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
patch()
Define a PATCH request
HttpRequest$patch( path = NULL, query = list(), body = NULL, disk = NULL, stream = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list
body
body as an R list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
encode
one of form, multipart, json, or raw
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
delete()
Define a DELETE request
HttpRequest$delete( path = NULL, query = list(), body = NULL, disk = NULL, stream = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list
body
body as an R list
disk
a path to write to. if NULL (default), memory used.
See curl::curl_fetch_disk()
for help.
stream
an R function to determine how to stream data. if
NULL (default), memory used. See curl::curl_fetch_stream()
for help
encode
one of form, multipart, json, or raw
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
head()
Define a HEAD request
HttpRequest$head(path = NULL, ...)
path
URL path, appended to the base URL
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
verb()
Use an arbitrary HTTP verb supported on this class Supported verbs: get, post, put, patch, delete, head
HttpRequest$verb(verb, ...)
verb
an HTTP verb supported on this class: get, post, put, patch, delete, head. Also supports retry.
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
z <- HttpRequest$new(url = "https://hb.opencpu.org/get") res <- z$verb('get', query = list(hello = "world")) res$payload
retry()
Define a RETRY request
HttpRequest$retry( verb, ..., pause_base = 1, pause_cap = 60, pause_min = 1, times = 3, terminate_on = NULL, retry_only_on = NULL, onwait = NULL )
verb
an HTTP verb supported on this class: get, post, put, patch, delete, head. Also supports retry.
...
curl options, only those in the acceptable set from
curl::curl_options()
except the following: httpget, httppost, post,
postfields, postfieldsize, and customrequest
pause_base, pause_cap, pause_min
basis, maximum, and minimum for
calculating wait time for retry. Wait time is calculated according to the
exponential backoff with full jitter algorithm. Specifically, wait time is
chosen randomly between pause_min
and the lesser of pause_base * 2
and
pause_cap
, with pause_base
doubling on each subsequent retry attempt.
Use pause_cap = Inf
to not terminate retrying due to cap of wait time
reached.
times
the maximum number of times to retry. Set to Inf
to
not stop retrying due to exhausting the number of attempts.
terminate_on, retry_only_on
a vector of HTTP status codes. For
terminate_on
, the status codes for which to terminate retrying, and for
retry_only_on
, the status codes for which to retry the request.
onwait
a callback function if the request will be retried and a wait time is being applied. The function will be passed two parameters, the response object from the failed request, and the wait time in seconds. Note that the time spent in the function effectively adds to the wait time, so it should be kept simple.
method()
Get the HTTP method (if defined)
HttpRequest$method()
(character) the HTTP method
clone()
The objects of this class are cloneable with this method.
HttpRequest$clone(deep = FALSE)
deep
Whether to make a deep clone.
Other async:
Async
,
AsyncQueue
,
AsyncVaried
## Not run: x <- HttpRequest$new(url = "https://hb.opencpu.org/get") ## note here how the HTTP method is shown on the first line to the right x$get() ## assign to a new object to keep the output z <- x$get() ### get the HTTP method z$method() (x <- HttpRequest$new(url = "https://hb.opencpu.org/get")$get()) x$url x$payload (x <- HttpRequest$new(url = "https://hb.opencpu.org/post")) x$post(body = list(foo = "bar")) HttpRequest$new( url = "https://hb.opencpu.org/get", headers = list( `Content-Type` = "application/json" ) ) # retry (x <- HttpRequest$new(url = "https://hb.opencpu.org/post")) x$retry("post", body = list(foo = "bar")) ## End(Not run) ## ------------------------------------------------ ## Method `HttpRequest$verb` ## ------------------------------------------------ z <- HttpRequest$new(url = "https://hb.opencpu.org/get") res <- z$verb('get', query = list(hello = "world")) res$payload
## Not run: x <- HttpRequest$new(url = "https://hb.opencpu.org/get") ## note here how the HTTP method is shown on the first line to the right x$get() ## assign to a new object to keep the output z <- x$get() ### get the HTTP method z$method() (x <- HttpRequest$new(url = "https://hb.opencpu.org/get")$get()) x$url x$payload (x <- HttpRequest$new(url = "https://hb.opencpu.org/post")) x$post(body = list(foo = "bar")) HttpRequest$new( url = "https://hb.opencpu.org/get", headers = list( `Content-Type` = "application/json" ) ) # retry (x <- HttpRequest$new(url = "https://hb.opencpu.org/post")) x$retry("post", body = list(foo = "bar")) ## End(Not run) ## ------------------------------------------------ ## Method `HttpRequest$verb` ## ------------------------------------------------ z <- HttpRequest$new(url = "https://hb.opencpu.org/get") res <- z$verb('get', query = list(hello = "world")) res$payload
Class with methods for handling HTTP responses
Additional Methods
raise_for_ct(type, charset = NULL, behavior = "stop")
Check response content-type; stop or warn if not matched. Parameters:
type: (character) a mime type to match against; see mime::mimemap for allowed values
charset: (character) if a charset string given, we check that it matches the charset in the content type header. default: NULL
behavior: (character) one of stop (default) or warning
raise_for_ct_html(charset = NULL, behavior = "stop")
Check that the response content-type is text/html
; stop or warn if
not matched. Parameters: see raise_for_ct()
raise_for_ct_json(charset = NULL, behavior = "stop")
Check that the response content-type is application/json
; stop or
warn if not matched. Parameters: see raise_for_ct()
raise_for_ct_xml(charset = NULL, behavior = "stop")
Check that the response content-type is application/xml
; stop or warn if
not matched. Parameters: see raise_for_ct()
This is an R6 class from the package R6. Find out more
about R6 at https://r6.r-lib.org/. After creating an instance of an R6
class (e.g., x <- HttpClient$new(url = "https://hb.opencpu.org")
) you can
access values and methods on the object x
.
method
(character) one or more URLs
url
(character) one or more URLs
opts
(character) one or more URLs
handle
(character) one or more URLs
status_code
(character) one or more URLs
request_headers
(character) one or more URLs
response_headers
(character) one or more URLs
response_headers_all
(character) one or more URLs
modified
(character) one or more URLs
times
(character) one or more URLs
content
(character) one or more URLs
request
(character) one or more URLs
raise_for_ct
for ct method (general)
raise_for_ct_html
for ct method (html)
raise_for_ct_json
for ct method (json)
raise_for_ct_xml
for ct method (xml)
print()
print method for HttpResponse objects
HttpResponse$print(x, ...)
x
self
...
ignored
new()
Create a new HttpResponse object
HttpResponse$new( method, url, opts, handle, status_code, request_headers, response_headers, response_headers_all, modified, times, content, request )
method
(character) HTTP method
url
(character) A url, required
opts
(list) curl options
handle
A handle
status_code
(integer) status code
request_headers
(list) request headers, named list
response_headers
(list) response headers, named list
response_headers_all
(list) all response headers, including intermediate redirect headers, unnamed list of named lists
modified
(character) modified date
times
(vector) named vector
content
(raw) raw binary content response
request
request object, with all details
parse()
Parse the raw response content to text
HttpResponse$parse(encoding = NULL, ...)
encoding
(character) A character string describing the
current encoding. If left as NULL
, we attempt to guess the
encoding. Passed to from
parameter in iconv
...
additional parameters passed on to iconv
(options: sub,
mark, toRaw). See ?iconv
for help
character string
success()
Was status code less than or equal to 201
HttpResponse$success()
boolean
status_http()
Get HTTP status code, message, and explanation
HttpResponse$status_http(verbose = FALSE)
verbose
(logical) whether to get verbose http status description,
default: FALSE
object of class "http_code", a list with slots for status_code, message, and explanation
raise_for_status()
Check HTTP status and stop with appropriate
HTTP error code and message if >= 300. otherwise use httpcode.
If you have fauxpas
installed we use that.
HttpResponse$raise_for_status()
stop or warn with message
clone()
The objects of this class are cloneable with this method.
HttpResponse$clone(deep = FALSE)
deep
Whether to make a deep clone.
## Not run: x <- HttpResponse$new(method = "get", url = "https://hb.opencpu.org") x$url x$method x <- HttpClient$new(url = 'https://hb.opencpu.org') (res <- x$get('get')) res$request_headers res$response_headers res$parse() res$status_code res$status_http() res$status_http()$status_code res$status_http()$message res$status_http()$explanation res$success() x <- HttpClient$new(url = 'https://hb.opencpu.org/status/404') (res <- x$get()) # res$raise_for_status() x <- HttpClient$new(url = 'https://hb.opencpu.org/status/414') (res <- x$get()) # res$raise_for_status() ## End(Not run)
## Not run: x <- HttpResponse$new(method = "get", url = "https://hb.opencpu.org") x$url x$method x <- HttpClient$new(url = 'https://hb.opencpu.org') (res <- x$get('get')) res$request_headers res$response_headers res$parse() res$status_code res$status_http() res$status_http()$status_code res$status_http()$message res$status_http()$explanation res$success() x <- HttpClient$new(url = 'https://hb.opencpu.org/status/404') (res <- x$get()) # res$raise_for_status() x <- HttpClient$new(url = 'https://hb.opencpu.org/status/414') (res <- x$get()) # res$raise_for_status() ## End(Not run)
Works for both synchronous requests via HttpClient()
and async
requests via Async()
and AsyncVaried()
mock(on = TRUE)
mock(on = TRUE)
on |
(logical) turn mocking on with |
webmockr
package required for mocking behavior
## Not run: if (interactive()) { library(webmockr) library(crul) URL <- "https://hb.opencpu.org" # turn on mocking crul::mock() # stub a request stub_request("get", file.path(URL, "get")) # create an HTTP client (x <- HttpClient$new(url = URL)) # make a request - matches stub - no real request made x$get('get') # allow net connect webmockr::webmockr_allow_net_connect() x$get('get', query = list(foo = "bar")) webmockr::webmockr_disable_net_connect() x$get('get', query = list(foo = "bar")) # With Async urls <- c( file.path(URL, "get"), file.path(URL, "anything"), file.path(URL, "encoding/utf8") ) for (u in urls) { webmockr::stub_request("get", u) %>% webmockr::to_return(body = list(mocked = TRUE)) } async_con <- Async$new(urls = urls) async_resp <- async_con$get() lapply(async_resp, \(x) x$parse("UTF-8")) } ## End(Not run)
## Not run: if (interactive()) { library(webmockr) library(crul) URL <- "https://hb.opencpu.org" # turn on mocking crul::mock() # stub a request stub_request("get", file.path(URL, "get")) # create an HTTP client (x <- HttpClient$new(url = URL)) # make a request - matches stub - no real request made x$get('get') # allow net connect webmockr::webmockr_allow_net_connect() x$get('get', query = list(foo = "bar")) webmockr::webmockr_disable_net_connect() x$get('get', query = list(foo = "bar")) # With Async urls <- c( file.path(URL, "get"), file.path(URL, "anything"), file.path(URL, "encoding/utf8") ) for (u in urls) { webmockr::stub_request("get", u) %>% webmockr::to_return(body = list(mocked = TRUE)) } async_con <- Async$new(urls = urls) async_resp <- async_con$get() lapply(async_resp, \(x) x$parse("UTF-8")) } ## End(Not run)
check if a url is okay
ok(x, status = 200L, info = TRUE, verb = "head", ua_random = FALSE, ...)
ok(x, status = 200L, info = TRUE, verb = "head", ua_random = FALSE, ...)
x |
either a URL as a character string, or an object of class HttpClient |
status |
(integer) one or more HTTP status codes, must be integers.
default: |
info |
(logical) in the case of an error, do you want a
|
verb |
(character) use "head" (default) or "get" HTTP verb for the request. note that "get" will take longer as it returns a body. however, "verb=get" may be your only option if a url blocks head requests |
ua_random |
(logical) use a random user agent string?
default: |
... |
args passed on to HttpClient |
We internally verify that status is an integer and in the known set of HTTP status codes, and that info is a boolean
You may have to fiddle with the parameters to ok()
as well as
curl options to get the "right answer". If you think you are
incorrectly getting FALSE
, the first thing to do is to pass in
verbose=TRUE
to ok()
. That will give you verbose curl output and will
help determine what the issue may be. Here's some different scenarios:
the site blocks head requests: some sites do this, try verb="get"
it will be hard to determine a site that requires this, but it's
worth trying a random useragent string, e.g., ok(useragent = "foobar")
some sites are up and reachable but you could get a 403 Unauthorized error, there's nothing you can do in this case other than having access
its possible to get a weird HTTP status code, e.g., LinkedIn gives a 999 code, they're trying to prevent any programmatic access
A FALSE
result may be incorrect depending on the use case. For example,
if you want to know if curl based scraping will work without fiddling with
curl options, then the FALSE
is probably correct, but if you want to
fiddle with curl options, then first step would be to send verbose=TRUE
to see whats going on with any redirects and headers. You can set headers,
user agent strings, etc. to get closer to the request you want to know
about. Note that a user agent string is always passed by default, but it
may not be the one you want.
a single boolean, if TRUE
the URL is up and okay,
if FALSE
it is down; but, see Details
## Not run: # 200 ok("https://www.google.com") # 200 ok("https://hb.opencpu.org/status/200") # more than one status ok("https://www.google.com", status = c(200L, 202L)) # 404 ok("https://hb.opencpu.org/status/404") # doesn't exist ok("https://stuff.bar") # doesn't exist ok("stuff") # use get verb instead of head ok("http://animalnexus.ca") ok("http://animalnexus.ca", verb = "get") # some urls will require a different useragent string # they probably regex the useragent string ok("https://doi.org/10.1093/chemse/bjq042") ok("https://doi.org/10.1093/chemse/bjq042", verb = "get", useragent = "foobar") # with random user agent's ## here, use a request hook to print out just the user agent string so ## we can see what user agent string is being sent off fun_ua <- function(request) { message(paste0("User-agent: ", request$options$useragent), sep = "\n") } z <- crul::HttpClient$new("https://doi.org/10.1093/chemse/bjq042", hooks = list(request = fun_ua)) z replicate(5, ok(z, ua_random=TRUE), simplify=FALSE) ## if you set useragent option it will override ua_random=TRUE ok("https://doi.org/10.1093/chemse/bjq042", useragent="foobar", ua_random=TRUE) # with HttpClient z <- crul::HttpClient$new("https://hb.opencpu.org/status/404", opts = list(verbose = TRUE)) ok(z) ## End(Not run)
## Not run: # 200 ok("https://www.google.com") # 200 ok("https://hb.opencpu.org/status/200") # more than one status ok("https://www.google.com", status = c(200L, 202L)) # 404 ok("https://hb.opencpu.org/status/404") # doesn't exist ok("https://stuff.bar") # doesn't exist ok("stuff") # use get verb instead of head ok("http://animalnexus.ca") ok("http://animalnexus.ca", verb = "get") # some urls will require a different useragent string # they probably regex the useragent string ok("https://doi.org/10.1093/chemse/bjq042") ok("https://doi.org/10.1093/chemse/bjq042", verb = "get", useragent = "foobar") # with random user agent's ## here, use a request hook to print out just the user agent string so ## we can see what user agent string is being sent off fun_ua <- function(request) { message(paste0("User-agent: ", request$options$useragent), sep = "\n") } z <- crul::HttpClient$new("https://doi.org/10.1093/chemse/bjq042", hooks = list(request = fun_ua)) z replicate(5, ok(z, ua_random=TRUE), simplify=FALSE) ## if you set useragent option it will override ua_random=TRUE ok("https://doi.org/10.1093/chemse/bjq042", useragent="foobar", ua_random=TRUE) # with HttpClient z <- crul::HttpClient$new("https://hb.opencpu.org/status/404", opts = list(verbose = TRUE)) ok(z) ## End(Not run)
A client to help you paginate
See HttpClient()
for information on parameters
a list, with objects of class HttpResponse()
.
Responses are returned in the order they are passed in.
This is an R6 class from the package R6. Find out more
about R6 at https://r6.r-lib.org/. After creating an instance of an R6
class (e.g., x <- HttpClient$new(url = "https://hb.opencpu.org")
) you can
access values and methods on the object x
.
Supported now:
limit_offset
: the most common way (in my experience), so is the default.
This method involves setting how many records and what record to start at
for each request. We send these query parameters for you.
page_perpage
: set the page to fetch and (optionally) how many records
to get per page
Supported later, hopefully:
link_headers
: link headers are URLS for the next/previous/last
request given in the response header from the server. This is relatively
uncommon, though is recommended by JSONAPI and is implemented by a
well known API (GitHub).
cursor
: this works by a single string given back in each response, to
be passed in the subsequent response, and so on until no more records
remain. This is common in Solr
http_req
an object of class HttpClient
by
(character) how to paginate. Only 'limit_offset' supported for now. In the future will support 'link_headers' and 'cursor'. See Details.
chunk
(numeric/integer) the number by which to chunk
requests, e.g., 10 would be be each request gets 10 records.
number is passed through format()
to prevent larger numbers
from being scientifically formatted
limit_param
(character) the name of the limit parameter. Default: limit
offset_param
(character) the name of the offset parameter. Default: offset
limit
(numeric/integer) the maximum records wanted.
number is passed through format()
to prevent larger numbers
from being scientifically formatted
page_param
(character) the name of the page parameter. Default: NULL
per_page_param
(character) the name of the per page parameter. Default: NULL
progress
(logical) print a progress bar, using utils::txtProgressBar.
Default: FALSE
.
print()
print method for Paginator
objects
Paginator$print(x, ...)
x
self
...
ignored
new()
Create a new Paginator
object
Paginator$new( client, by = "limit_offset", limit_param = NULL, offset_param = NULL, limit = NULL, chunk = NULL, page_param = NULL, per_page_param = NULL, progress = FALSE )
client
an object of class HttpClient
, from a call to HttpClient
by
(character) how to paginate. Only 'limit_offset' supported for now. In the future will support 'link_headers' and 'cursor'. See Details.
limit_param
(character) the name of the limit parameter. Default: limit
offset_param
(character) the name of the offset parameter. Default: offset
limit
(numeric/integer) the maximum records wanted
chunk
(numeric/integer) the number by which to chunk requests, e.g., 10 would be be each request gets 10 records
page_param
(character) the name of the page parameter.
per_page_param
(character) the name of the per page parameter.
progress
(logical) print a progress bar, using utils::txtProgressBar.
Default: FALSE
.
A new Paginator
object
get()
make a paginated GET request
Paginator$get(path = NULL, query = list(), ...)
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
post()
make a paginated POST request
Paginator$post( path = NULL, query = list(), body = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
body
body as an R list
encode
one of form, multipart, json, or raw
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
put()
make a paginated PUT request
Paginator$put( path = NULL, query = list(), body = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
body
body as an R list
encode
one of form, multipart, json, or raw
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
patch()
make a paginated PATCH request
Paginator$patch( path = NULL, query = list(), body = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
body
body as an R list
encode
one of form, multipart, json, or raw
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
delete()
make a paginated DELETE request
Paginator$delete( path = NULL, query = list(), body = NULL, encode = "multipart", ... )
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
body
body as an R list
encode
one of form, multipart, json, or raw
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
head()
make a paginated HEAD request
Paginator$head(path = NULL, ...)
path
URL path, appended to the base URL
...
For retry
, the options to be passed on to the method
implementing the requested verb, including curl options. Otherwise,
curl options, only those in the acceptable set from curl::curl_options()
except the following: httpget, httppost, post, postfields, postfieldsize,
and customrequest
not sure if this makes any sense or not yet
responses()
list responses
Paginator$responses()
a list of HttpResponse
objects, empty list before requests made
status_code()
Get HTTP status codes for each response
Paginator$status_code()
numeric vector, empty numeric vector before requests made
status()
List HTTP status objects
Paginator$status()
a list of http_code
objects, empty list before requests made
parse()
parse content
Paginator$parse(encoding = "UTF-8")
encoding
(character) the encoding to use in parsing. default:"UTF-8"
character vector, empty character vector before requests made
content()
Get raw content for each response
Paginator$content()
raw list, empty list before requests made
times()
curl request times
Paginator$times()
list of named numeric vectors, empty list before requests made
url_fetch()
get the URL that would be sent (i.e., before executing the request) the only things that change the URL are path and query parameters; body and any curl options don't change the URL
Paginator$url_fetch(path = NULL, query = list())
path
URL path, appended to the base URL
query
query terms, as a named list. any numeric values are
passed through format()
to prevent larger numbers from being
scientifically formatted
URLs (character)
\dontrun{ cli <- HttpClient$new(url = "https://api.crossref.org") cc <- Paginator$new(client = cli, limit_param = "rows", offset_param = "offset", limit = 50, chunk = 10) cc$url_fetch('works') cc$url_fetch('works', query = list(query = "NSF")) }
clone()
The objects of this class are cloneable with this method.
Paginator$clone(deep = FALSE)
deep
Whether to make a deep clone.
## Not run: if (interactive()) { # limit/offset approach con <- HttpClient$new(url = "https://api.crossref.org") cc <- Paginator$new(client = con, limit_param = "rows", offset_param = "offset", limit = 50, chunk = 10) cc cc$get('works') cc cc$responses() cc$status() cc$status_code() cc$times() # cc$content() cc$parse() lapply(cc$parse(), jsonlite::fromJSON) # page/per page approach (with no per_page param allowed) conn <- HttpClient$new(url = "https://discuss.ropensci.org") cc <- Paginator$new(client = conn, by = "page_perpage", page_param = "page", per_page_param = "per_page", limit = 90, chunk = 30) cc cc$get('c/usecases/l/latest.json') cc$responses() lapply(cc$parse(), jsonlite::fromJSON) # page/per_page conn <- HttpClient$new('https://api.inaturalist.org') cc <- Paginator$new(conn, by = "page_perpage", page_param = "page", per_page_param = "per_page", limit = 90, chunk = 30) cc cc$get('v1/observations', query = list(taxon_name="Helianthus")) cc$responses() res <- lapply(cc$parse(), jsonlite::fromJSON) res[[1]]$total_results vapply(res, "[[", 1L, "page") vapply(res, "[[", 1L, "per_page") vapply(res, function(w) NROW(w$results), 1L) ## another ccc <- Paginator$new(conn, by = "page_perpage", page_param = "page", per_page_param = "per_page", limit = 500, chunk = 30, progress = TRUE) ccc ccc$get('v1/observations', query = list(taxon_name="Helianthus")) res2 <- lapply(ccc$parse(), jsonlite::fromJSON) vapply(res2, function(w) NROW(w$results), 1L) # progress bar (con <- HttpClient$new(url = "https://api.crossref.org")) cc <- Paginator$new(client = con, limit_param = "rows", offset_param = "offset", limit = 50, chunk = 10, progress = TRUE) cc cc$get('works') } ## End(Not run) ## ------------------------------------------------ ## Method `Paginator$url_fetch` ## ------------------------------------------------ ## Not run: cli <- HttpClient$new(url = "https://api.crossref.org") cc <- Paginator$new(client = cli, limit_param = "rows", offset_param = "offset", limit = 50, chunk = 10) cc$url_fetch('works') cc$url_fetch('works', query = list(query = "NSF")) ## End(Not run)
## Not run: if (interactive()) { # limit/offset approach con <- HttpClient$new(url = "https://api.crossref.org") cc <- Paginator$new(client = con, limit_param = "rows", offset_param = "offset", limit = 50, chunk = 10) cc cc$get('works') cc cc$responses() cc$status() cc$status_code() cc$times() # cc$content() cc$parse() lapply(cc$parse(), jsonlite::fromJSON) # page/per page approach (with no per_page param allowed) conn <- HttpClient$new(url = "https://discuss.ropensci.org") cc <- Paginator$new(client = conn, by = "page_perpage", page_param = "page", per_page_param = "per_page", limit = 90, chunk = 30) cc cc$get('c/usecases/l/latest.json') cc$responses() lapply(cc$parse(), jsonlite::fromJSON) # page/per_page conn <- HttpClient$new('https://api.inaturalist.org') cc <- Paginator$new(conn, by = "page_perpage", page_param = "page", per_page_param = "per_page", limit = 90, chunk = 30) cc cc$get('v1/observations', query = list(taxon_name="Helianthus")) cc$responses() res <- lapply(cc$parse(), jsonlite::fromJSON) res[[1]]$total_results vapply(res, "[[", 1L, "page") vapply(res, "[[", 1L, "per_page") vapply(res, function(w) NROW(w$results), 1L) ## another ccc <- Paginator$new(conn, by = "page_perpage", page_param = "page", per_page_param = "per_page", limit = 500, chunk = 30, progress = TRUE) ccc ccc$get('v1/observations', query = list(taxon_name="Helianthus")) res2 <- lapply(ccc$parse(), jsonlite::fromJSON) vapply(res2, function(w) NROW(w$results), 1L) # progress bar (con <- HttpClient$new(url = "https://api.crossref.org")) cc <- Paginator$new(client = con, limit_param = "rows", offset_param = "offset", limit = 50, chunk = 10, progress = TRUE) cc cc$get('works') } ## End(Not run) ## ------------------------------------------------ ## Method `Paginator$url_fetch` ## ------------------------------------------------ ## Not run: cli <- HttpClient$new(url = "https://api.crossref.org") cc <- Paginator$new(client = cli, limit_param = "rows", offset_param = "offset", limit = 50, chunk = 10) cc$url_fetch('works') cc$url_fetch('works', query = list(query = "NSF")) ## End(Not run)
progress bars
pass httr::progress()
to progress
param in HttpClient,
which pulls out relevant info to pass down to curl
if file sizes known you get progress bar; if file sizes not known you get bytes downloaded
See the README for examples
proxy options
proxy(url, user = NULL, pwd = NULL, auth = "basic")
proxy(url, user = NULL, pwd = NULL, auth = "basic")
url |
(character) URL, with scheme (http/https), domain and port (must be numeric). required. |
user |
(character) username, optional |
pwd |
(character) password, optional |
auth |
(character) authentication type, one of basic (default),
digest, digest_ie, gssnegotiate, ntlm, any or |
See https://www.hidemyass.com/proxy for a list of proxies you can use
proxy("http://97.77.104.22:3128") proxy("97.77.104.22:3128") proxy("http://97.77.104.22:3128", "foo", "bar") proxy("http://97.77.104.22:3128", "foo", "bar", auth = "digest") proxy("http://97.77.104.22:3128", "foo", "bar", auth = "ntlm") # socks proxy("socks5://localhost:9050/", auth = NULL) ## Not run: # with proxy (look at request/outgoing headers) # (res <- HttpClient$new( # url = "http://www.google.com", # proxies = proxy("http://97.77.104.22:3128") # )) # res$proxies # res$get(verbose = TRUE) # vs. without proxy (look at request/outgoing headers) # (res2 <- HttpClient$new(url = "http://www.google.com")) # res2$get(verbose = TRUE) # Use authentication # (res <- HttpClient$new( # url = "http://google.com", # proxies = proxy("http://97.77.104.22:3128", user = "foo", pwd = "bar") # )) # another example # (res <- HttpClient$new( # url = "http://ip.tyk.nu/", # proxies = proxy("http://200.29.191.149:3128") # )) # res$get()$parse("UTF-8") ## End(Not run)
proxy("http://97.77.104.22:3128") proxy("97.77.104.22:3128") proxy("http://97.77.104.22:3128", "foo", "bar") proxy("http://97.77.104.22:3128", "foo", "bar", auth = "digest") proxy("http://97.77.104.22:3128", "foo", "bar", auth = "ntlm") # socks proxy("socks5://localhost:9050/", auth = NULL) ## Not run: # with proxy (look at request/outgoing headers) # (res <- HttpClient$new( # url = "http://www.google.com", # proxies = proxy("http://97.77.104.22:3128") # )) # res$proxies # res$get(verbose = TRUE) # vs. without proxy (look at request/outgoing headers) # (res2 <- HttpClient$new(url = "http://www.google.com")) # res2$get(verbose = TRUE) # Use authentication # (res <- HttpClient$new( # url = "http://google.com", # proxies = proxy("http://97.77.104.22:3128", user = "foo", pwd = "bar") # )) # another example # (res <- HttpClient$new( # url = "http://ip.tyk.nu/", # proxies = proxy("http://200.29.191.149:3128") # )) # res$get()$parse("UTF-8") ## End(Not run)
upload file
upload(path, type = NULL)
upload(path, type = NULL)
path |
(character) a single path, file must exist |
type |
(character) a file type, guessed by mime::guess_type if not given |
## Not run: # image path <- file.path(Sys.getenv("R_DOC_DIR"), "html/logo.jpg") (x <- HttpClient$new(url = "https://hb.opencpu.org")) res <- x$post(path = "post", body = list(y = upload(path))) res$content # text file, in a list file <- upload(system.file("CITATION")) res <- x$post(path = "post", body = list(y = file)) jsonlite::fromJSON(res$parse("UTF-8")) # text file, as data res <- x$post(path = "post", body = file) jsonlite::fromJSON(res$parse("UTF-8")) ## End(Not run)
## Not run: # image path <- file.path(Sys.getenv("R_DOC_DIR"), "html/logo.jpg") (x <- HttpClient$new(url = "https://hb.opencpu.org")) res <- x$post(path = "post", body = list(y = upload(path))) res$content # text file, in a list file <- upload(system.file("CITATION")) res <- x$post(path = "post", body = list(y = file)) jsonlite::fromJSON(res$parse("UTF-8")) # text file, as data res <- x$post(path = "post", body = file) jsonlite::fromJSON(res$parse("UTF-8")) ## End(Not run)
Build and parse URLs
url_build(url, path = NULL, query = NULL) url_parse(url)
url_build(url, path = NULL, query = NULL) url_parse(url)
url |
(character) a url, length 1 |
path |
(character) a path, length 1 |
query |
(list) a named list of query parameters |
url_build
returns a character string URL; url_parse
returns a list with URL components
url_build("https://hb.opencpu.org") url_build("https://hb.opencpu.org", "get") url_build("https://hb.opencpu.org", "post") url_build("https://hb.opencpu.org", "get", list(foo = "bar")) url_parse("hb.opencpu.org") url_parse("https://hb.opencpu.org") url_parse(url = "https://hb.opencpu.org") url_parse("https://hb.opencpu.org/get") url_parse("https://hb.opencpu.org/get?foo=bar") url_parse("https://hb.opencpu.org/get?foo=bar&stuff=things") url_parse("https://hb.opencpu.org/get?foo=bar&stuff=things[]")
url_build("https://hb.opencpu.org") url_build("https://hb.opencpu.org", "get") url_build("https://hb.opencpu.org", "post") url_build("https://hb.opencpu.org", "get", list(foo = "bar")) url_parse("hb.opencpu.org") url_parse("https://hb.opencpu.org") url_parse(url = "https://hb.opencpu.org") url_parse("https://hb.opencpu.org/get") url_parse("https://hb.opencpu.org/get?foo=bar") url_parse("https://hb.opencpu.org/get?foo=bar&stuff=things") url_parse("https://hb.opencpu.org/get?foo=bar&stuff=things[]")
The DELETE method deletes the specified resource.
The DELETE method requests that the origin server remove the association between the target resource and its current functionality. In effect, this method is similar to the rm command in UNIX: it expresses a deletion operation on the URI mapping of the origin server rather than an expectation that the previously associated information be deleted.
https://datatracker.ietf.org/doc/html/rfc7231#section-4.3.5
Other verbs:
verb-GET
,
verb-HEAD
,
verb-PATCH
,
verb-POST
,
verb-PUT
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$delete(path = 'delete') ## a list (res1 <- x$delete('delete', body = list(hello = "world"), verbose = TRUE)) jsonlite::fromJSON(res1$parse("UTF-8")) ## a string (res2 <- x$delete('delete', body = "hello world", verbose = TRUE)) jsonlite::fromJSON(res2$parse("UTF-8")) ## empty body request x$delete('delete', verbose = TRUE) ## End(Not run)
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$delete(path = 'delete') ## a list (res1 <- x$delete('delete', body = list(hello = "world"), verbose = TRUE)) jsonlite::fromJSON(res1$parse("UTF-8")) ## a string (res2 <- x$delete('delete', body = "hello world", verbose = TRUE)) jsonlite::fromJSON(res2$parse("UTF-8")) ## empty body request x$delete('delete', verbose = TRUE) ## End(Not run)
The GET method requests a representation of the specified resource. Requests using GET should only retrieve data.
The GET method requests transfer of a current selected representation for the target resource. GET is the primary mechanism of information retrieval and the focus of almost all performance optimizations. Hence, when people speak of retrieving some identifiable information via HTTP, they are generally referring to making a GET request.
It is tempting to think of resource identifiers as remote file system pathnames and of representations as being a copy of the contents of such files. In fact, that is how many resources are implemented (see Section 9.1 (https://datatracker.ietf.org/doc/html/rfc7231#section-9.1) for related security considerations). However, there are no such limitations in practice. The HTTP interface for a resource is just as likely to be implemented as a tree of content objects, a programmatic view on various database records, or a gateway to other information systems. Even when the URI mapping mechanism is tied to a file system, an origin server might be configured to execute the files with the request as input and send the output as the representation rather than transfer the files directly. Regardless, only the origin server needs to know how each of its resource identifiers corresponds to an implementation and how each implementation manages to select and send a current representation of the target resource in a response to GET.
A client can alter the semantics of GET to be a "range request", requesting transfer of only some part(s) of the selected representation, by sending a Range header field in the request (RFC7233: https://datatracker.ietf.org/doc/html/rfc7233).
A payload within a GET request message has no defined semantics; sending a payload body on a GET request might cause some existing implementations to reject the request.
The response to a GET request is cacheable; a cache MAY use it to satisfy subsequent GET and HEAD requests unless otherwise indicated by the Cache-Control header field (Section 5.2 of RFC7234: https://datatracker.ietf.org/doc/html/rfc7234#section-5.2).
https://datatracker.ietf.org/doc/html/rfc7231#section-4.3.1
Other verbs:
verb-DELETE
,
verb-HEAD
,
verb-PATCH
,
verb-POST
,
verb-PUT
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$get(path = 'get') ## End(Not run)
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$get(path = 'get') ## End(Not run)
The HEAD method asks for a response identical to that of a GET request, but without the response body.
The HEAD method is identical to GET except that the server MUST NOT send a message body in the response (i.e., the response terminates at the end of the header section). The server SHOULD send the same header fields in response to a HEAD request as it would have sent if the request had been a GET, except that the payload header fields MAY be omitted. This method can be used for obtaining metadata about the selected representation without transferring the representation data and is often used for testing hypertext links for validity, accessibility, and recent modification.
https://datatracker.ietf.org/doc/html/rfc7231#section-4.3.2
Other verbs:
verb-DELETE
,
verb-GET
,
verb-PATCH
,
verb-POST
,
verb-PUT
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$head() ## End(Not run)
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$head() ## End(Not run)
The PATCH method is used to apply partial modifications to a resource.
The PATCH method requests that a set of changes described in the request entity be applied to the resource identified by the Request- URI. The set of changes is represented in a format called a "patch document" identified by a media type. If the Request-URI does not point to an existing resource, the server MAY create a new resource, depending on the patch document type (whether it can logically modify a null resource) and permissions, etc.
https://datatracker.ietf.org/doc/html/rfc5789
Other verbs:
verb-DELETE
,
verb-GET
,
verb-HEAD
,
verb-POST
,
verb-PUT
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$patch(path = 'patch', body = list(hello = "mars")) ## End(Not run)
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$patch(path = 'patch', body = list(hello = "mars")) ## End(Not run)
The POST method is used to submit an entity to the specified resource, often causing a change in state or side effects on the server.
If one or more resources has been created on the origin server as a result of successfully processing a POST request, the origin server SHOULD send a 201 (Created) response containing a Location header field that provides an identifier for the primary resource created (Section 7.1.2 https://datatracker.ietf.org/doc/html/rfc7231#section-7.1.2) and a representation that describes the status of the request while referring to the new resource(s).
https://datatracker.ietf.org/doc/html/rfc7231#section-4.3.3
Other verbs:
verb-DELETE
,
verb-GET
,
verb-HEAD
,
verb-PATCH
,
verb-PUT
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") # a named list x$post(path='post', body = list(hello = "world")) # a string x$post(path='post', body = "hello world") # an empty body request x$post(path='post') # encode="form" res <- x$post(path="post", encode = "form", body = list( custname = 'Jane', custtel = '444-4444', size = 'small', topping = 'bacon', comments = 'make it snappy' ) ) jsonlite::fromJSON(res$parse("UTF-8")) # encode="json" res <- x$post("post", encode = "json", body = list( genus = 'Gagea', species = 'pratensis' ) ) jsonlite::fromJSON(res$parse()) ## End(Not run)
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") # a named list x$post(path='post', body = list(hello = "world")) # a string x$post(path='post', body = "hello world") # an empty body request x$post(path='post') # encode="form" res <- x$post(path="post", encode = "form", body = list( custname = 'Jane', custtel = '444-4444', size = 'small', topping = 'bacon', comments = 'make it snappy' ) ) jsonlite::fromJSON(res$parse("UTF-8")) # encode="json" res <- x$post("post", encode = "json", body = list( genus = 'Gagea', species = 'pratensis' ) ) jsonlite::fromJSON(res$parse()) ## End(Not run)
The PUT method replaces all current representations of the target resource with the request payload.
The PUT method requests that the state of the target resource be created or replaced with the state defined by the representation enclosed in the request message payload. A successful PUT of a given representation would suggest that a subsequent GET on that same target resource will result in an equivalent representation being sent in a 200 (OK) response. However, there is no guarantee that such a state change will be observable, since the target resource might be acted upon by other user agents in parallel, or might be subject to dynamic processing by the origin server, before any subsequent GET is received. A successful response only implies that the user agent's intent was achieved at the time of its processing by the origin server.
If the target resource does not have a current representation and the PUT successfully creates one, then the origin server MUST inform the user agent by sending a 201 (Created) response. If the target resource does have a current representation and that representation is successfully modified in accordance with the state of the enclosed representation, then the origin server MUST send either a 200 (OK) or a 204 (No Content) response to indicate successful completion of the request.
https://datatracker.ietf.org/doc/html/rfc7231#section-4.3.4
Other verbs:
verb-DELETE
,
verb-GET
,
verb-HEAD
,
verb-PATCH
,
verb-POST
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$put(path = 'put', body = list(foo = "bar")) ## End(Not run)
## Not run: x <- HttpClient$new(url = "https://hb.opencpu.org") x$put(path = 'put', body = list(foo = "bar")) ## End(Not run)
Writing data options
## Not run: # write to disk (x <- HttpClient$new(url = "https://hb.opencpu.org")) f <- tempfile() res <- x$get("get", disk = f) res$content # when using write to disk, content is a path readLines(res$content) close(file(f)) # streaming response (x <- HttpClient$new(url = "https://hb.opencpu.org")) res <- x$get('stream/50', stream = function(x) cat(rawToChar(x))) res$content # when streaming, content is NULL ## Async (cc <- Async$new( urls = c( 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar', 'https://hb.opencpu.org/get?b=4', 'https://hb.opencpu.org/get?stuff=things', 'https://hb.opencpu.org/get?b=4&g=7&u=9&z=1' ) )) files <- replicate(5, tempfile()) (res <- cc$get(disk = files, verbose = TRUE)) lapply(files, readLines) ## Async varied ### disk f <- tempfile() g <- tempfile() req1 <- HttpRequest$new(url = "https://hb.opencpu.org/get")$get(disk = f) req2 <- HttpRequest$new(url = "https://hb.opencpu.org/post")$post(disk = g) req3 <- HttpRequest$new(url = "https://hb.opencpu.org/get")$get() (out <- AsyncVaried$new(req1, req2, req3)) out$request() out$content() readLines(f) readLines(g) out$parse() close(file(f)) close(file(g)) ### stream - to console fun <- function(x) print(x) req1 <- HttpRequest$new(url = "https://hb.opencpu.org/get" )$get(query = list(foo = "bar"), stream = fun) req2 <- HttpRequest$new(url = "https://hb.opencpu.org/get" )$get(query = list(hello = "world"), stream = fun) (out <- AsyncVaried$new(req1, req2)) out$request() out$content() ### stream - to an R object lst <- list() fun <- function(x) lst <<- append(lst, list(x)) req1 <- HttpRequest$new(url = "https://hb.opencpu.org/get" )$get(query = list(foo = "bar"), stream = fun) req2 <- HttpRequest$new(url = "https://hb.opencpu.org/get" )$get(query = list(hello = "world"), stream = fun) (out <- AsyncVaried$new(req1, req2)) out$request() lst cat(vapply(lst, function(z) rawToChar(z$content), ""), sep = "\n") ## End(Not run)
## Not run: # write to disk (x <- HttpClient$new(url = "https://hb.opencpu.org")) f <- tempfile() res <- x$get("get", disk = f) res$content # when using write to disk, content is a path readLines(res$content) close(file(f)) # streaming response (x <- HttpClient$new(url = "https://hb.opencpu.org")) res <- x$get('stream/50', stream = function(x) cat(rawToChar(x))) res$content # when streaming, content is NULL ## Async (cc <- Async$new( urls = c( 'https://hb.opencpu.org/get?a=5', 'https://hb.opencpu.org/get?foo=bar', 'https://hb.opencpu.org/get?b=4', 'https://hb.opencpu.org/get?stuff=things', 'https://hb.opencpu.org/get?b=4&g=7&u=9&z=1' ) )) files <- replicate(5, tempfile()) (res <- cc$get(disk = files, verbose = TRUE)) lapply(files, readLines) ## Async varied ### disk f <- tempfile() g <- tempfile() req1 <- HttpRequest$new(url = "https://hb.opencpu.org/get")$get(disk = f) req2 <- HttpRequest$new(url = "https://hb.opencpu.org/post")$post(disk = g) req3 <- HttpRequest$new(url = "https://hb.opencpu.org/get")$get() (out <- AsyncVaried$new(req1, req2, req3)) out$request() out$content() readLines(f) readLines(g) out$parse() close(file(f)) close(file(g)) ### stream - to console fun <- function(x) print(x) req1 <- HttpRequest$new(url = "https://hb.opencpu.org/get" )$get(query = list(foo = "bar"), stream = fun) req2 <- HttpRequest$new(url = "https://hb.opencpu.org/get" )$get(query = list(hello = "world"), stream = fun) (out <- AsyncVaried$new(req1, req2)) out$request() out$content() ### stream - to an R object lst <- list() fun <- function(x) lst <<- append(lst, list(x)) req1 <- HttpRequest$new(url = "https://hb.opencpu.org/get" )$get(query = list(foo = "bar"), stream = fun) req2 <- HttpRequest$new(url = "https://hb.opencpu.org/get" )$get(query = list(hello = "world"), stream = fun) (out <- AsyncVaried$new(req1, req2)) out$request() lst cat(vapply(lst, function(z) rawToChar(z$content), ""), sep = "\n") ## End(Not run)