Ecosyste.ms: Awesome

An open API service indexing awesome lists of open source software.

Awesome Lists | Featured Topics | Projects

https://github.com/hrbrmstr/curlconverter

:curly_loop: :arrow_right: :heavy_minus_sign: Translate cURL command lines into parameters for use with httr or actual httr calls (R)
https://github.com/hrbrmstr/curlconverter

curl httr r r-cyber rstats

Last synced: 7 days ago
JSON representation

:curly_loop: :arrow_right: :heavy_minus_sign: Translate cURL command lines into parameters for use with httr or actual httr calls (R)

Awesome Lists containing this project

README

        

---
output: rmarkdown::github_document
editor_options:
chunk_output_type: console
---

[![Build Status](https://travis-ci.org/hrbrmstr/curlconverter.svg?branch=master)](https://travis-ci.org/hrbrmstr/curlconverter)
[![AppVeyor Build Status](https://ci.appveyor.com/api/projects/status/github/hrbrmstrcurlconverter?branch=master&svg=true)](https://ci.appveyor.com/project/hrbrmstr/curlconverter)
[![CRAN_Status_Badge](https://www.r-pkg.org/badges/version/curlconverter)](https://cran.r-project.org/package=curlconverter)
[![](http://cranlogs.r-pkg.org/badges/curlconverter)](https://cran.r-project.org/package=curlconverter)
[![Coverage Status](https://img.shields.io/codecov/c/github///master.svg)](https://codecov.io/github//?branch=master)

# curlconverter

Tools to Transform 'cURL' Command-Line Calls to 'httr' Requests

## Description

Deciphering web/'REST' 'API' and 'XHR' calls can be tricky, which is one reason why internet browsers provide "Copy as cURL" functionality within their "Developer Tools"pane(s). These 'cURL' command-lines can be difficult to wrangle into an 'httr' 'GET' or 'POST' request, but you can now "straighten" these 'cURLs' either from data copied to the system clipboard or by passing in a vector of 'cURL' command-lines and getting back a list of parameter elements which can be used to form 'httr' requests. These lists can be passed to another function to automagically make 'httr' functions.

### WIP

This is the path back to CRAN and to 1.0.0.

The V8 dependency has been removed and the package now uses `docopt`. This will make it easier to install on many systems and enable rapid addition of support for new `cURL` command-line parameters.

The "HAR" functions are not working well with the new methods but will be for full release.

## What's Inside The Tin

The following functions are implemented:

- `straighten`: convert one or more _"Copy as cURL"_ command lines into useful data
- `parse_query`: parse URL query parameters into a named list
- `make_req`: turn parsed cURL command lines into a `httr` request functions (i.e. returns working R functions)

The following functions are implemented:

## Installation

```{r eval=FALSE}
devtools::install_github("hrbrmstr/curlconverter")
```

```{r message=FALSE, warning=FALSE, error=FALSE, include=FALSE}
options(width=120)
```

## Usage

```{r message=FALSE, warning=FALSE, error=FALSE}
library(curlconverter)
library(jsonlite)
library(httr)

# current verison
packageVersion("curlconverter")
```

Simple example using a call to :

```{r}
httpbinrhcurl <- "curl 'https://httpbin.org/headers' -H 'pragma: no-cache' -H 'accept-encoding: gzip, deflate, sdch' -H 'accept-language: en-US,en;q=0.8' -H 'upgrade-insecure-requests: 1' -H 'user-agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/49.0.2623.39 Safari/537.36' -H 'accept: text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8' -H 'cache-control: no-cache' -H 'referer: https://httpbin.org/' --compressed"

straight <- straighten(httpbinrhcurl)
res <- make_req(straight)

# or

straighten(httpbinrhcurl) %>%
make_req() -> res

toJSON(content(res[[1]](), as="parsed"), auto_unbox = TRUE, pretty=TRUE)
```

Slightly more complex one:

```{r}
toJSON(straighten("curl 'http://financials.morningstar.com/ajax/ReportProcess4HtmlAjax.html?&t=XNAS:MSFT&region=usa&culture=en-US&cur=&reportType=is&period=12&dataType=A&order=asc&columnYear=5&curYearPart=1st5year&rounding=3&view=raw&r=973302&callback=jsonp1454021128757&_=1454021129337' -H 'Cookie: JSESSIONID=5E43C98903E865D72AA3C2DCEF317848; sfhabit=asc%7Craw%7C3%7C12%7CA%7C5%7Cv0.14; ScrollY=0' -H 'DNT: 1' -H 'Accept-Encoding: gzip, deflate, sdch' -H 'Accept-Language: en-US,en;q=0.8' -H 'User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_11_3) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/47.0.2526.111 Safari/537.36' -H 'Accept: text/javascript, application/javascript, */*' -H 'Referer: http://financials.morningstar.com/income-statement/is.html?t=MSFT&region=usa&culture=en-US' -H 'X-Requested-With: XMLHttpRequest' -H 'Connection: keep-alive' -H 'Cache-Control: max-age=0' --compressed"), auto_unbox = TRUE, pretty=TRUE)
```

There are some built-in test files you can play with:

```{r}
(curl_line <- readLines(system.file("extdata/curl5.txt", package="curlconverter"), warn=FALSE))

toJSON(straighten(curl_line, quiet=TRUE), auto_unbox = TRUE, pretty=TRUE)
(curl_line <- readLines(system.file("extdata/curl8.txt", package="curlconverter"), warn=FALSE))

# example with query parameters in the body
req <- straighten(curl_line, quiet=FALSE)

# ugh
(req[[1]]$data)

#yay!
toJSON(parse_query(req[[1]]$data), auto_unbox = TRUE, pretty=TRUE)
```

Spinning straw into gold

```{r}
curl_line <- c('curl "http://anasim.iet.unipi.it/moniqa/php/from_js.php" -H "Origin: http://anasim.iet.unipi.it" -H "Accept-Encoding: gzip, deflate" -H "Accept-Language: it-IT,it;q=0.8,en-US;q=0.6,en;q=0.4" -H "User-Agent: Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/48.0.2564.103 Safari/537.36" -H "Content-Type: application/x-www-form-urlencoded; charset=UTF-8" -H "Accept: */*" -H "Referer: http://anasim.iet.unipi.it/moniqa/" -H "X-Requested-With: XMLHttpRequest" -H "Connection: keep-alive" --data "deviceid=65&function_name=extract_measurements" --compressed')

straighten(curl_line) %>%
make_req() -> get_data

cat(capture.output(get_data[[1]]), sep="\n")
```

That also sends this to the console and clipboard:

```{r eval=FALSE}
httr::VERB(
verb = "GET", url = "http://anasim.iet.unipi.it/moniqa/php/from_js.php",
httr::add_headers(
Origin = "http://anasim.iet.unipi.it",
`Accept-Encoding` = "gzip, deflate",
`Accept-Language` = "it-IT,it;q=0.8,en-US;q=0.6,en;q=0.4",
`User-Agent` = "Mozilla/5.0 (Windows NT 6.3; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/48.0.2564.103 Safari/537.36",
Accept = "*/*", Referer = "http://anasim.iet.unipi.it/moniqa/",
`X-Requested-With` = "XMLHttpRequest",
Connection = "keep-alive"
),
body = list(
deviceid = "65",
function_name = "extract_measurements"
),
encode = "form"
)
```

### Code of Conduct

Please note that this project is released with a [Contributor Code of Conduct](CONDUCT.md).

By participating in this project you agree to abide by its terms.