Package: arrow 22.0.0

arrow: Integration to 'Apache' 'Arrow'
'Apache' 'Arrow' <https://arrow.apache.org/> is a cross-language development platform for in-memory data. It specifies a standardized language-independent columnar memory format for flat and hierarchical data, organized for efficient analytic operations on modern hardware. This package provides an interface to the 'Arrow C++' library.
Authors:
arrow_22.0.0.tar.gz
arrow_22.0.0.zip(r-4.6)arrow_22.0.0.zip(r-4.5)arrow_22.0.0.zip(r-4.4)
arrow_22.0.0.tgz(r-4.5-x86_64)arrow_22.0.0.tgz(r-4.5-arm64)arrow_22.0.0.tgz(r-4.4-x86_64)arrow_22.0.0.tgz(r-4.4-arm64)
arrow_22.0.0.tar.gz(r-4.6-arm64)arrow_22.0.0.tar.gz(r-4.6-x86_64)arrow_22.0.0.tar.gz(r-4.5-arm64)arrow_22.0.0.tar.gz(r-4.5-x86_64)
arrow.pdf |arrow.html✨
arrow/json (API)
NEWS
| # Install 'arrow' in R: |
| install.packages('arrow', repos = c('https://staging.r-multiverse.org', 'https://cloud.r-project.org')) |
Bug tracker:https://github.com/apache/arrow/issues
Pkgdown/docs site:https://arrow.apache.org
- bzip2– High-quality block-sorting file compressor library
- brotli– Library implementing brotli encoder and decoder
- zlib– Compression library
- lz4– Fast LZ compression algorithm library
- libzstd– Fast lossless compression algorithm
- curl– Easy-to-use client-side URL transfer library
- openssl– Secure Sockets Layer toolkit
- c++– GNU Standard C++ Library v3
arrowparquetbzip2brotlizliblz4libzstdcurlopensslcpp
Last updated from:5aeb5f217f (on apache-arrow-22.0.0). Checks:8 OK, 4 NOTE, 1 FAIL. Indexed: no.
| Target | Result | Total time | Artifact |
|---|---|---|---|
| linux-devel-arm64 | OK | 1709 | |
| linux-devel-x86_64 | OK | 1750 | |
| source / vignettes | OK | 1364 | |
| linux-release-arm64 | OK | 1355 | |
| linux-release-x86_64 | OK | 1366 | |
| macos-release-arm64 | OK | 475 | |
| macos-release-x86_64 | OK | 542 | |
| macos-oldrel-arm64 | NOTE | 329 | |
| macos-oldrel-x86_64 | NOTE | 1020 | |
| windows-devel | NOTE | 1499 | |
| windows-release | OK | 753 | |
| windows-oldrel | NOTE | 765 | |
| wasm-release | FAIL | 127 |
Exports:all_ofArrayarrow_arrayarrow_availablearrow_infoarrow_tablearrow_with_aceroarrow_with_datasetarrow_with_gcsarrow_with_jsonarrow_with_parquetarrow_with_s3arrow_with_substraitas_arrow_arrayas_arrow_tableas_chunked_arrayas_data_typeas_record_batchas_record_batch_readeras_schemabinaryboolbooleanbufferBufferBufferOutputStreamBufferReadercall_functioncast_optionschunked_arrayChunkedArrayCodeccodec_is_availableCompressedInputStreamCompressedOutputStreamCompressionTypeconcat_arraysconcat_tablescontainscopy_filescpu_countcreate_package_with_all_dependenciescsv_convert_optionscsv_parse_optionscsv_read_optionscsv_write_optionsCsvConvertOptionsCsvFileFormatCsvFragmentScanOptionsCsvParseOptionsCsvReadOptionsCsvTableReaderCsvWriteOptionsDatasetdataset_factoryDatasetFactorydate32date64DateUnitdecimaldecimal128decimal256decimal32decimal64default_memory_pooldictionaryDictionaryArrayDirectoryPartitioningDirectoryPartitioningFactorydurationends_witheverythingExpressionExtensionArrayExtensionTypeFeatherReaderfieldFieldFileFormatFileInfoFileModeFileOutputStreamFileSelectorFileSystemFileSystemDatasetFileSystemDatasetFactoryFileTypefixed_size_binaryfixed_size_list_ofFixedSizeListArrayFixedSizeListTypeflight_connectflight_disconnectflight_getflight_path_existsflight_putfloatfloat16float32float64FragmentScanOptionsGcsFileSystemgs_buckethalffloathive_partitionHivePartitioningHivePartitioningFactoryinfer_schemainfer_typeInMemoryDatasetinstall_arrowinstall_pyarrowint16int32int64int8io_thread_countIpcFileFormatis_inJoinTypeJsonFileFormatJsonFragmentScanOptionsJsonParseOptionsJsonReadOptionsJsonTableReaderlarge_binarylarge_list_oflarge_utf8LargeListArraylast_collist_compute_functionslist_flightslist_ofListArrayload_flight_serverLocalFileSystemmap_batchesmap_ofMapArrayMapTypematch_arrowmatchesMemoryMappedFileMessageReaderMessageTypeMetadataVersionmmap_createmmap_opennew_extension_arraynew_extension_typenullNullEncodingBehaviorNullHandlingBehaviornum_rangeone_ofopen_csv_datasetopen_datasetopen_delim_datasetopen_tsv_datasetParquetArrowReaderPropertiesParquetFileFormatParquetFileReaderParquetFileWriterParquetFragmentScanOptionsParquetReaderPropertiesParquetVersionTypeParquetWriterPropertiesPartitioningQuantileInterpolationRandomAccessFileread_csv_arrowread_csv2_arrowread_delim_arrowread_featherread_ipc_fileread_ipc_streamread_json_arrowread_messageread_parquetread_schemaread_tsv_arrowReadableFilerecord_batchRecordBatchRecordBatchFileReaderRecordBatchFileWriterRecordBatchReaderRecordBatchStreamReaderRecordBatchStreamWriterregister_extension_typeregister_scalar_functionreregister_extension_typeRoundModes3_bucketS3FileSystemscalarScalarScannerScannerBuilderschemaSchemaset_cpu_countset_io_thread_countshow_exec_planstarts_withStatusCodestringstructStructArrayStructScalarSubTreeFileSystemTabletime32time64timestampTimestampParserTimeUnitto_arrowto_duckdbtypeTypeuint16uint32uint64uint8unify_schemasUnionDatasetunregister_extension_typeutf8value_countsvctrs_extension_arrayvctrs_extension_typewrite_csv_arrowwrite_csv_datasetwrite_datasetwrite_delim_datasetwrite_featherwrite_ipc_filewrite_ipc_streamwrite_parquetwrite_to_rawwrite_tsv_dataset
Dependencies:assertthatbitbit64clicpp11gluelifecyclemagrittrpurrrR6rlangtidyselectvctrswithr
Readme and manuals
Help Manual
| Help page | Topics |
|---|---|
| Functions available in Arrow dplyr queries | acero arrow-dplyr arrow-functions arrow-verbs |
| Array Classes | Array DictionaryArray FixedSizeListArray LargeListArray ListArray MapArray StructArray |
| ArrayData class | ArrayData |
| Create an Arrow Array | arrow_array |
| Report information on the package's capabilities | arrow_available arrow_info arrow_with_acero arrow_with_dataset arrow_with_gcs arrow_with_json arrow_with_parquet arrow_with_s3 arrow_with_substrait |
| Create an Arrow Table | arrow_table |
| Convert an object to an Arrow Array | as_arrow_array as_arrow_array.Array as_arrow_array.ChunkedArray as_arrow_array.Scalar |
| Convert an object to an Arrow Table | as_arrow_table as_arrow_table.arrow_dplyr_query as_arrow_table.data.frame as_arrow_table.Dataset as_arrow_table.default as_arrow_table.RecordBatch as_arrow_table.RecordBatchReader as_arrow_table.Schema as_arrow_table.Table |
| Convert an object to an Arrow ChunkedArray | as_chunked_array as_chunked_array.Array as_chunked_array.ChunkedArray |
| Convert an object to an Arrow DataType | as_data_type as_data_type.DataType as_data_type.Field as_data_type.Schema |
| Convert an object to an Arrow RecordBatch | as_record_batch as_record_batch.arrow_dplyr_query as_record_batch.data.frame as_record_batch.RecordBatch as_record_batch.Table |
| Convert an object to an Arrow RecordBatchReader | as_record_batch_reader as_record_batch_reader.arrow_dplyr_query as_record_batch_reader.data.frame as_record_batch_reader.Dataset as_record_batch_reader.function as_record_batch_reader.RecordBatch as_record_batch_reader.RecordBatchReader as_record_batch_reader.Scanner as_record_batch_reader.Table |
| Convert an object to an Arrow Schema | as_schema as_schema.Schema as_schema.StructType |
| Create a Buffer | buffer |
| Buffer class | Buffer |
| Call an Arrow compute function | call_function |
| Create a Chunked Array | chunked_array |
| ChunkedArray class | ChunkedArray |
| Compression Codec class | Codec |
| Check whether a compression codec is available | codec_is_available |
| Compressed stream classes | CompressedInputStream CompressedOutputStream compression |
| Concatenate zero or more Arrays | c.Array concat_arrays |
| Concatenate one or more Tables | concat_tables |
| Copy files between FileSystems | copy_files |
| Manage the global CPU thread pool in libarrow | cpu_count set_cpu_count |
| Create a source bundle that includes all thirdparty dependencies | create_package_with_all_dependencies |
| CSV Convert Options | csv_convert_options |
| CSV Parsing Options | csv_parse_options |
| CSV Reading Options | csv_read_options |
| CSV Writing Options | csv_write_options |
| CSV dataset file format | CsvFileFormat |
| File reader options | CsvConvertOptions CsvParseOptions CsvReadOptions CsvWriteOptions JsonParseOptions JsonReadOptions TimestampParser |
| Arrow CSV and JSON table reader classes | CsvTableReader JsonTableReader |
| Create Arrow data types | binary bool boolean data-type date32 date64 decimal decimal128 decimal256 decimal32 decimal64 duration FixedSizeListType fixed_size_binary fixed_size_list_of float float16 float32 float64 halffloat int16 int32 int64 int8 large_binary large_list_of large_utf8 list_of MapType map_of null string struct time32 time64 timestamp uint16 uint32 uint64 uint8 utf8 |
| Multi-file datasets | Dataset DatasetFactory FileSystemDataset FileSystemDatasetFactory InMemoryDataset UnionDataset |
| Create a DatasetFactory | dataset_factory |
| DataType class | DataType |
| Create a dictionary type | dictionary |
| class DictionaryType | DictionaryType |
| Arrow expressions | Expression |
| ExtensionArray class | ExtensionArray |
| ExtensionType class | ExtensionType |
| FeatherReader class | FeatherReader |
| Create a Field | field |
| Field class | Field |
| Dataset file formats | FileFormat IpcFileFormat ParquetFileFormat |
| FileSystem entry info | FileInfo |
| file selector | FileSelector |
| FileSystem classes | FileSystem GcsFileSystem LocalFileSystem S3FileSystem SubTreeFileSystem |
| Format-specific write options | FileWriteOptions |
| FixedWidthType class | FixedWidthType |
| Connect to a Flight server | flight_connect |
| Explicitly close a Flight client | flight_disconnect |
| Get data from a Flight server | flight_get |
| Send data to a Flight server | flight_put |
| Format-specific scan options | CsvFragmentScanOptions FragmentScanOptions JsonFragmentScanOptions ParquetFragmentScanOptions |
| Connect to a Google Cloud Storage (GCS) bucket | gs_bucket |
| Construct Hive partitioning | hive_partition |
| Extract a schema from an object | infer_schema |
| Infer the arrow Array type from an R object | infer_type type |
| InputStream classes | BufferReader InputStream MemoryMappedFile RandomAccessFile ReadableFile |
| Install or upgrade the Arrow library | install_arrow |
| Install pyarrow for use with reticulate | install_pyarrow |
| Manage the global I/O thread pool in libarrow | io_thread_count set_io_thread_count |
| JSON dataset file format | JsonFileFormat |
| List available Arrow C++ compute functions | list_compute_functions |
| See available resources on a Flight server | flight_path_exists list_flights |
| Load a Python Flight server | load_flight_server |
| Apply a function to a stream of RecordBatches | map_batches |
| Value matching for Arrow objects | is_in match_arrow |
| Message class | Message |
| MessageReader class | MessageReader |
| Create a new read/write memory mapped file of a given size | mmap_create |
| Open a memory mapped file | mmap_open |
| Extension types | new_extension_array new_extension_type register_extension_type reregister_extension_type unregister_extension_type |
| Open a multi-file dataset | open_dataset |
| Open a multi-file dataset of CSV or other delimiter-separated format | open_csv_dataset open_delim_dataset open_tsv_dataset |
| OutputStream classes | BufferOutputStream FileOutputStream OutputStream |
| ParquetArrowReaderProperties class | ParquetArrowReaderProperties |
| ParquetFileReader class | ParquetFileReader |
| ParquetFileWriter class | ParquetFileWriter |
| ParquetReaderProperties class | ParquetReaderProperties |
| ParquetWriterProperties class | ParquetWriterProperties |
| Define Partitioning for a Dataset | DirectoryPartitioning DirectoryPartitioningFactory HivePartitioning HivePartitioningFactory Partitioning |
| Read a CSV or other delimited file with Arrow | read_csv2_arrow read_csv_arrow read_delim_arrow read_tsv_arrow |
| Read a Feather file (an Arrow IPC file) | read_feather read_ipc_file |
| Read Arrow IPC stream format | read_ipc_stream |
| Read a JSON file | read_json_arrow |
| Read a Message from a stream | read_message |
| Read a Parquet file | read_parquet |
| Read a Schema from a stream | read_schema |
| Create a RecordBatch | record_batch |
| RecordBatch class | RecordBatch |
| RecordBatchReader classes | RecordBatchFileReader RecordBatchReader RecordBatchStreamReader |
| RecordBatchWriter classes | RecordBatchFileWriter RecordBatchStreamWriter RecordBatchWriter |
| Register user-defined functions | register_scalar_function |
| Connect to an AWS S3 bucket | s3_bucket |
| Create an Arrow Scalar | scalar StructScalar |
| Arrow scalars | Scalar |
| Scan the contents of a dataset | Scanner ScannerBuilder |
| Create a schema or extract one from an object. | schema |
| Schema class | Schema |
| Show the details of an Arrow Execution Plan | show_exec_plan |
| Table class | Table |
| Create an Arrow object from a DuckDB connection | to_arrow |
| Create a (virtual) DuckDB table from an Arrow object | to_duckdb |
| Combine and harmonize schemas | unify_schemas |
| 'table' for Arrow objects | value_counts |
| Extension type for generic typed vectors | vctrs_extension_array vctrs_extension_type |
| Write CSV file to disk | write_csv_arrow |
| Write a dataset | write_dataset |
| Write a dataset into partitioned flat files. | write_csv_dataset write_delim_dataset write_tsv_dataset |
| Write a Feather file (an Arrow IPC file) | write_feather write_ipc_file |
| Write Arrow IPC stream format | write_ipc_stream |
| Write Parquet file to disk | write_parquet |
| Write Arrow data to a raw vector | write_to_raw |
