Raku Land

The uploading author of cpan:ANTONOV does not match the META author of github:antononcube.

Data::ExampleDatasets

cpan:ANTONOV

Data::ExampleDatasets Raku package

Raku package for (obtaining) example datasets.

Currently, this repository contains only datasets metadata. The datasets are downloaded from the repository Rdatasets, [VAB1].


Usage examples

Setup

Here we load the Raku modules Data::Generators, Data::Summarizers, and this module, Data::ExampleDatasets:

use Data::Reshapers;
use Data::Summarizers;
use Data::ExampleDatasets;
# (Any)

Get a dataset by using an identifier

Here we get a dataset by using an identifier and display part of the obtained dataset:

my @tbl = example-dataset('Baumann', :headers);
say to-pretty-table(@tbl[^6]);
#ERROR: If the first argument is an array then it is expected that it can be coerced into a array-of-hashes, array-of-positionals, or hash-of-hashes, which in turn can be coerced into a full two dimensional array.
# Nil

Here we summarize the dataset obtained above:

records-summary(@tbl)
#ERROR: No such method 'value' for invocant of type 'Array'.  Did you mean
#ERROR: 'values'?
# Nil

Remark: The values for the first argument of example-dataset correspond to the values of the columns “Item” and “Package”, respectively, in the metadata dataset from the GitHub repository “Rdatasets”, [VAB1]. See the datasets metadata sub-section below.

The first argument of example-dataset can take as values:

Get a dataset by using an URL

Here we get a dataset by using an URL and display a summary of the obtained dataset:

my $url = 'https://raw.githubusercontent.com/antononcube/Raku-Data-Reshapers/main/resources/dfTitanic.csv';
my @tbl2 = example-dataset($url, :headers);
records-summary(@tbl2);
# +-----------------+----------------+---------------------+-------------------+---------------+
# | id              | passengerClass | passengerAge        | passengerSurvival | passengerSex  |
# +-----------------+----------------+---------------------+-------------------+---------------+
# | Min    => 1     | 3rd => 709     | Min    => -1        | died     => 809   | male   => 843 |
# | 1st-Qu => 327.5 | 1st => 323     | 1st-Qu => 10        | survived => 500   | female => 466 |
# | Mean   => 655   | 2nd => 277     | Mean   => 23.550038 |                   |               |
# | Median => 655   |                | Median => 20        |                   |               |
# | 3rd-Qu => 982.5 |                | 3rd-Qu => 40        |                   |               |
# | Max    => 1309  |                | Max    => 80        |                   |               |
# +-----------------+----------------+---------------------+-------------------+---------------+

Datasets metadata

Here we:

  1. Get the dataset of the datasets metadata
  2. Filter it to have only datasets with 13 rows
  3. Keep only the columns "Item", "Title", "Rows", and "Cols"
  4. Display it in "pretty table" format
my @tblMeta = get-datasets-metadata();
@tblMeta = @tblMeta.grep({ $_<Rows> == 13}).map({ $_.grep({ $_.key (elem) <Item Title Rows Cols>}).Hash });
say to-pretty-table(@tblMeta)
# +------+------------+------+--------------------------------------------------------------------+
# | Cols |    Item    | Rows |                               Title                                |
# +------+------------+------+--------------------------------------------------------------------+
# |  4   | Snow.pumps |  13  |    John Snow's Map and Data on the 1854 London Cholera Outbreak    |
# |  7   |    BCG     |  13  |                          BCG Vaccine Data                          |
# |  5   |   cement   |  13  |                  Heat Evolved by Setting Cements                   |
# |  2   |  kootenay  |  13  |   Waterflow Measurements of Kootenay River in Libby and Newgate    |
# |  5   | Newhouse77 |  13  | Medical-Care Expenditure: A Cross-National Survey (Newhouse, 1977) |
# |  2   |   Saxony   |  13  |                         Families in Saxony                         |
# +------+------------+------+--------------------------------------------------------------------+

Keeping downloaded data

By default the data is obtained over the web from Rdatasets, but example-dataset has an option to keep the data "locally." (The data is saved in XDG_DATA_HOME, see [JS1].)

This can be demonstrated with the following timings of a dataset with ~1300 rows:

my $startTime = now;
my $data = example-dataset( / 'COUNT::titanic' $ / ):keep;
my $endTime = now;
say "Geting the data first time took { $endTime - $startTime } seconds";
# Geting the data first time took 2.90335873 seconds
$startTime = now;
$data = example-dataset( / 'COUNT::titanic' $/ ):keep;
$endTime = now;
say "Geting the data second time took { $endTime - $startTime } seconds";
# Geting the data second time took 1.559416817 seconds

References

Functions, packages, repositories

[AAf1] Anton Antonov, ExampleDataset, (2020), Wolfram Function Repository.

[VAB1] Vincent Arel-Bundock, Rdatasets, (2020), GitHub/vincentarelbundock.

[JS1] Jonathan Stowe, XDG::BaseDirectory, (last updated on 2021-03-31), Raku Modules.

Interactive interfaces

[AAi1] Anton Antonov, Example datasets recommender interface, (2021), Shinyapps.io.