A python package to read and write R RData and Rds files into/from
pandas dataframes. It does not need to have R or other external
It can read mainly R data frames and tibbles. Also supports vectors, matrices, arrays and tables.
R lists and R S4 objects (such as those from Bioconductor) are not supported. Please read the
Known limitations section and the section on what objects can be read for more information.
Detailed documentation on all available methods is in the Module documentation
Moving from R to Python and fighting against indentation issues? Missing curly braces? Missing the <- operator for assignment? Then try PytwisteR! Python with a twist of R! (note: it works, but it's only a joke)
The package depends on pandas, which you normally have installed if you got Anaconda (highly recommended.) If creating a new conda or virtual environment or if you don't have it in your base installation, pandas should get installed automatically.
If you are reading 3D arrays, you will need to install xarray manually. This is not installed automatically as most users won't need it.
In order to compile from source, you will need a C compiler (see installation) and cython (version >= 0.28).
librdata also depends on zlib, bzip2 and lzma; it was reported not to be installed on Lubuntu or docker base ubuntu images. If you face this problem intalling the libraries solves it.
Probably the easiest way: from your conda, virtualenv or just base installation do:
pip install pyreadr
If you are running on a machine without admin rights, and you want to install against your base installation you can do:
pip install pyreadr --user
We offer pre-compiled wheels for Windows, linux and macOs.
The package is also available in conda-forge for windows, mac and linux 64 bit.
In order to install:
conda install -c conda-forge pyreadr
Download or clone the repo, open a command window and type:
python3 setup.py install
If you don't have admin privileges to the machine do:
python3 setup.py install --user
You can also install from the github repo directly (without cloning). Use the flag --user if necessary.
pip install git+https://github.com/ofajardo/pyreadr.git
You need a working C compiler and cython.
Pass the path to a RData or Rds file to the function read_r. It will return a dictionary with object names as keys and pandas data frames as values.
For example, in order to read a RData file:
import pyreadr result = pyreadr.read_r('test_data/basic/two.RData') # done! let's see what we got print(result.keys()) # let's check what objects we got df1 = result["df1"] # extract the pandas data frame for object df1
reading a Rds file is equally simple. Rds files have one single object, which you can access with the key None:
import pyreadr result = pyreadr.read_r('test_data/basic/one.Rds') # done! let's see what we got print(result.keys()) # let's check what objects we got: there is only None df1 = result[None] # extract the pandas data frame for the only object available
Here there is a relation of all functions available. You can also check the Module documentation.
|Function in this package||Purpose|
|read_r||reads RData and Rds files|
|list_objects||list objects and column names contained in RData or Rds file|
|download_file||download file from internet|
|write_rdata||writes RData files|
|write_rds||writes Rds files|
Pyreadr allows you to write one single pandas data frame into a single R dataframe and store it into a RData or Rds file. Other python or R object types are not supported. Writing more than one object is not supported.
import pyreadr import pandas as pd # prepare a pandas dataframe df = pd.DataFrame([["a",1],["b",2]], columns=["A", "B"]) # let's write into RData # df_name is the name for the dataframe in R, by default dataset pyreadr.write_rdata("test.RData", df, df_name="dataset") # now let's write a Rds pyreadr.write_rds("test.Rds", df) # done!
now you can check the result in R:
load("test.RData") print(dataset) dataset2 <- readRDS("test.Rds") print(dataset2)
By default the resulting files will be uncompressed, you can activate gzip compression by passing the option compress="gzip". This is useful in case you have big files.
import pyreadr import pandas as pd # prepare a pandas dataframe df = pd.DataFrame([["a",1],["b",2]], columns=["A", "B"]) # write a compressed RData file pyreadr.write_rdata("test.RData", df, df_name="dataset", compress="gzip") # write a compressed Rds file pyreadr.write_rds("test.Rds", df, compress="gzip")
Librdata, the C backend of pyreadr absolutely needs a file in disk and only a string with the path can be passed as argument, therefore you cannot pass an url to pyreadr.read_r.
In order to help with this limitation, pyreadr provides a funtion download_file which as its name suggests downloads a file from an url to disk:
import pyreadr url = "https://github.com/hadley/nycflights13/blob/master/data/airlines.rda?raw=true" dst_path = "/some/path/on/disk/airlines.rda" dst_path_again = pyreadr.download_file(url, dst_path) res = pyreadr.read_r(dst_path)
As you see download_file returns the path where the file was written, therefore you can pass it to pyreadr.read_r directly:
import pyreadr url = "https://github.com/hadley/nycflights13/blob/master/data/airlines.rda?raw=true" dst_path = "/some/path/on/disk/airlines.rda" res = pyreadr.read_r(pyreadr.download_file(url, dst_path), dst_path)
You can use the argument use_objects of the function read_r to specify which objects should be read.
import pyreadr result = pyreadr.read_r('test_data/basic/two.RData', use_objects=["df1"]) # done! let's see what we got print(result.keys()) # let's check what objects we got, now only df1 is listed df1 = result["df1"] # extract the pandas data frame for object df1
The function list_objects gives a dictionary with object names contained in the RData or Rds file as keys and a list of column names as values. It is not always possible to retrieve column names without reading the whole file in those cases you would get None instead of a column name.
import pyreadr object_list = pyreadr.list_objects('test_data/basic/two.RData') # done! let's see what we got print(object_list) # let's check what objects we got and what columns those have
R Date objects are read as datetime.date objects.
R datetime objects (POSIXct and POSIXlt) are internally stored as UTC timestamps, and may have additional timezone information if the user set it explicitly. If no timezone information was set by the user R uses the local timezone for display.
librdata cannot retrieve that timezone information, therefore pyreadr display UTC time by default, which will not match the display in R. You can set explicitly some timezone (your local timezone for example) with the argument timezone for the function read_r
import pyreadr result = pyreadr.read_r('test_data/basic/two.RData', timezone='CET')
if you would like to just use your local timezone as R does, you can get it with tzlocal (you need to install it first with pip) and pass the information to read_r:
import tzlocal import pyreadr my_timezone = tzlocal.get_localzone().zone result = pyreadr.read_r('test_data/basic/two.RData', timezone=my_timezone)
If you have control over the data in R, a good option to avoid all of this is to transform the POSIX object to character, then transform it to a datetime in python.
When writing these kind of objects pyreadr transforms them to characters. Those can be easily transformed back to POSIX with as.POSIXct/lt (see later).
Data frames composed of character, numeric (double), integer, timestamp (POSIXct and POSIXlt), date, logical atomic vectors. Factors are also supported.
Tibbles are also supported.
Atomic vectors as described before can also be directly read and are translated to a pandas data frame with one column.
Matrices, arrays and tables are also read and translated to pandas data frames (because those objects in R can be named, and plain numpy arrays do not support dimension names). The only exception is 3D arrays, which are translated to a xarray DataArray (as pandas does not support more than 2 dimensions). This is also the only time that an object different from a pandas dataframe is returned by read_r.
For 3D arrays, consider that python prints these in a different way as R does, but still you are looking at the same array (see for example here for an explanation.)
Only single pandas data frames can be written into R data frames.
Lists and S4 objects (such as those coming from Bioconductor are not supported. Please read the Known limitations section for more information.
For converting python/numpy types to R types the following rules are followed:
|Python Type||R Type|
|np.int32 or lower||integer|
|category||depends on the original dtype|
|any other object||character|
|column all missing||logical|
|column with mixed types||character|
datetime and date objects are translated to character to avoid problems with timezones. These characters can be easily translated back to POSIXct/lt in R using as.POSIXct/lt. The format of the datetimes/dates is prepared for this but can be controlled with the arguments dateformat and datetimeformat for write_rdata and write_rds. Those arguments take python standard formatting strings.
Pandas categories are NOT translated to R factors. Instead the original data type of the category is preserved and transformed according to the rules. This is because R factors are integers and levels are always strings, in pandas factors can be any type and leves any type as well, therefore it is not always adecquate to coerce everything to the integer/character system. In the other hand, pandas category level information is lost in the process.
Any other object is transformed to a character using the str representation of the object.
Columns with mixed types are translated to character. This does not apply to column cotaining np.nan, where the missing values are correctly translated.
R integers are 32 bit. Therefore python 64 bit integer have to be promoted to numeric in order to fit.
A pandas column containing only missing values is transformed to logical, following R's behavior.
librdata writes Numeric missing values as NaN instead of NA. In pandas we only have np.nan both as NaN and missing value representation, and it will always be written as NaN in R.
POSIXct and POSIXlt objects in R are stored internally as UTC timestamps and may have in addition time zone information. librdata does not return time zone information and thefore the display of the tiemstamps in R and in pandas may differ.
Librdata reads arrays with a maximum of 3 dimensions. If more dimensions are present you will get an error. Please submit an issue if this is the case.
Lists are not read.
S4 Objects and probably other kind of objects, including those that depend on non base R packages (Bioconductor for example) cannot be read. The error code in this case is as follows:
"pyreadr.custom_errors.LibrdataError: The file contains an unrecognized object"
Data frames with special values like arrays, matrices and other data frames are not supported.
librdata first de-compresses the file in memory and then extracts the data. That means you need more free RAM than the decompress file ocuppies in memory. RData and Rds files are highly compressed: they can occupy in memory easily 40 or even more times in memory as in disk. Take it into account in case you get a "Unable to allocate memory" error (see this )
When writing numeric missing values are translated to NaN instead of NA.
Writing is supported only for a single pandas data frame to a single R data frame. Other data types are not supported. Multiple data frames for rdata files are not supported.
RData and Rds files produced by R are (by default) compressed. Files produced by pyreadr are not compressed by default and therefore pretty bulky in comparison. You can pass the option compress="gzip" to write_rds or write_rda in order to activate gzip compression.
Pyreadr writing is a relative slow operation compared to doint it in R.
Cannot read RData or rds files in encodings other than utf-8.
Solutions to some of these limitations have been proposed in the upstream librdata issues (points 1-4 are addressed by issue 12, point 5 by issue 16 and point 7 by issue 17). However there is no guarantee that these changes will be made and there are no timelines either. If you think it would be nice if these issues are solved, please express your support in the librdata issues.
Contributions are welcome! Please chech the document CONTRIBUTING.md for more details.
A log with the changes for each version can be found here
Otto Fajardo - author, maintainer
Jonathon Love - contributor (original cython wrapper from jamovi-readstat and msvc compatible librdata)
deenes - reading lzma compression
Daniel M. Sullivan - added license information to setup.py