Data Central Ingestion Documentation

Data Central simplifies astronomical data releases by automating the ingestion, versioning, web-based data access, and hosting of your data and supporting documentation.

Data Central has many features for hosting astronomical data including:

  • Image cutout service produces publication-quality single-band or rgb images with custom image scaling options, fits creation, overplotting sources from DC-hosted and remote catalogues, batch mode, contours and more, over 30 bands across the GAMA and DEVILS regions (with VHS data coming soon)

  • Single object viewer including image cutout at source position, data product download, 2D map visualizer, tabular data

  • Cone search service including remote source position overlays

  • Auto-generated, interactive, intuitive schema browser

  • SQL/ADQL query service; return records that match some criteria from the catalogues hosted at Data Central

  • TAP service

  • Authenticated users can re-visit their query/cutout/cone search history

Your data will be discoverable alongside many of Australia’s national optical surveys. Data Central also provides access to our in-house CMS, Document Central, for your survey team to curate the data-release’s supporting documentation.

Tip

Data Central hosts a few different types of documentation:

  • static documentation on a specific data release (appearing in the schema browser)

  • updatable documentation about a survey (Document Central)

  • team wikis

  • documentation on how to use Data Central itself (including how to cite Data Central and the surveys we host).

Only the first of these is discussed in the ingestion documentation, and only the first of these needs to be provided by the survey/data team before ingestion can begin. The others can be changed at any time, outside of the constraints of the ingestion process.

Think of the first as a paper—you cannot change it once you’ve published it, you can only publish a new paper with the corrections (allowing citations)—whereas the others are like wikipedia, in that the content can be continuously updated, and hence is a poor source for citation.

Getting Started

Data Central can currently ingest, link, cross-match and serve spectra, images, catalogues and IFS data products. We have detailed descriptions for how to format your data depending on its type, but first, follow the instructions below to setup the basic file structure and metadata files for your data release.

Metadata files are provided alongside your data and allow us to populate the Schema Browser to best describe your data. These files are un-editable once your data has been released via Data Central (prior to the final sign-off from the survey team, we will go through an iterative process with you to verify that the data are correct and displayed as expected. During that stage you are free to make changes to metadata files).

Documentation intended for public consumption that you will want to update frequently (e.g., with detailed descriptions of analysis or version update information) should be maintained by the survey teams themselves in Data Central’s Documentation portal: Documentation Central. Contact us to get your team set up within the system if you have not done so already.

Whilst we can be flexible about how we receive the data and metadata, here’s what we see as the quickest and easiest way to get your data ingested:

  1. Create a Data Central Account at https://accounts.datacentral.org.au/register/ (if you have not already done so).

  2. Request a team at https://teams.datacentral.org.au/request-team/ (if you do not already have a team for the survey/data). The team will need to be approved by a Data Central admin.

  3. Log in with your Data Central account at https://dev.aao.org.au/ (using the Data Central button), as this will allow you git access to the repository which stores all the metadata. If multiple members of the team want to be updating the metadata, have them also create Data Central accounts and login to https://dev.aao.org.au, and include their usernames in a ticket here https://jira.aao.org.au/servicedesk/customer/portal/3/create/28.

  4. Upload your data in the correct layout (as noted below) to https://cloud.datacentral.org.au/.

  5. Make a merge request on https://dev.aao.org.au/ with the metadata as documented below.

  6. Check the data in the pre-release link we (the Data Central team) send you, and sign off on the release.

Directory Structure

To ingest data into Data Central, you will provide two folders, one containing the data products themselves, and one containing the metadata. The data/ and metadata/ directories will be further populated according to the types of data products you wish to release (see section Data Types for specific requirements; catalogue/ifs/spectra etc).

Data

A top-level <survey> directory should contain a single directory per <datarelease>.

data
└── <survey>
    └── <datarelease>

Metadata

The following file structure should be adopted. A top-level <survey> directory should contain a single directory per <datarelease>. Both directories should have metadata files described below which will populate the Schema Browser.

metadata
└── <survey>
    ├── <survey>_survey_meta.txt
    └── <datarelease>
        └── <survey>_<data_release>_data_release_meta.txt

Metadata Files

Attention

Metadata files are always pipe-delimited, and have the extension .txt

There are two files you’ll need to provide to initialize your survey and data release in Data Central, <survey>_survey_meta.txt and <survey>_<datarelease>_data_release_meta.txt.

<survey>_survey_meta.txt

This file describes your survey at a high level, and is used to populate the Schema Browser. It is uneditable once the data are released, detailed documentation that is editable by the teams should be written into a Document Central article (contact us if you do not yet have an account).

Provide the following a single pipe-delimited .txt file containing an entry (row) for your survey:

name

pretty_name

title

description

pi

contact

website

gama

GAMA

Galaxy and Mass Assembly Survey

GAMA is a project to exploit the latest generation of ground-based and space-borne survey facilities to study cosmology and galaxy formation and evolution. At the heart of this project lies the GAMA spectroscopic survey of ~300,000 galaxies down to r < 19.8 mag over ~286 deg2, carried out using the AAOmega multi-object spectrograph on the Anglo-Australian Telescope (AAT) by the GAMA team. This project was awarded 210 nights over 7 years (2008–2014) and the observations are now completed. This survey builds on, and is augmented by, previous spectroscopic surveys such as the Sloan Digital Sky Survey (SDSS), the 2dF Galaxy Redshift Survey(2dFGRS) and the Millennium Galaxy Catalogue (MGC).

Simon Driver

Simon Driver

http://www/gama-survey.org/

Replace <survey> in the filename with the value for name in the first column.

<survey>_survey_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

A machine-readable, lowercase version of your survey name (e.g., GAMA would be entered as gama). Use only alphanumeric characters (no spaces and do not start your survey with a digit). This must be unique within Data Central (we cannot support multiple surveys named ‘DEVILS’, you will need to think of a new name!). This should match the string used in the file name.

pretty_name(required=True, type=char, max_limit=100)

A human-readable version of the survey name. This can contain any characters (up to the character limit).

title(required=True, type=char, max_limit=100)

A longer version of the survey name (likely an expansion of the acronym!)

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing your survey.

pi(required=True, type=char, max_limit=100)

The name of the Principal Investigator

contact(required=True, type=char, max_limit=100)

Format as: John Smith <john.smith@institute.org>

website(required=True, type=char, max_limit=500)

The survey team’s website for public consumption (e.g., https://devilsurvey.org/)

Note

If you have previously hosted data with Data Central, you can skip providing the <survey>_survey_meta.txt file. We’ll already have this from your previous data release.

<survey>_<datarelease>_data_release_meta.txt

This file describes your data release at a high level, and is used to populate the Schema Browser. It is uneditable once the data are released, detailed documentation that is editable by the teams should be written into a Document Central article (contact us if you do not yet have an account).

Provide the following a single pipe-delimited .txt file containing an entry (row) for your data release:

name

pretty_name

version

data_release_number

contact

dr2

Data Release 2

1

2

Simon Driver <simon.driver@uwa.edu.au>

Replace <datarelease> in the filename with the value for name in the first column.

<survey>_<datarelease>_data_release_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

A machine-readable, lowercase version of your data release name (e.g., Data Release 2 would be entered as dr2). Use only alphanumeric characters (no spaces). This must be unique within your survey (you cannot release dr2 twice!). This should match the string used for the data release in the file name.

pretty_name(required=True, type=char, max_limit=100)

A human-readable version of the data release name. This can contain any characters (up to the character limit).

version(required=True, type=float, default=1.0)

The version of the data release data. This is meaningful to the team only, and should be described in the team-curated and created Documentation

data_release_number(required=True, type=int, default=1)

The number of the data release within Data Central. e.g., if this data release is named “Final Data Release”, which may correspond to the 5th data release, we need a numeric representation here so as to show the data releases in the schema browser sequentially.

contact(required=True, type=char, max_limit=100)

Format as: John Smith <john.smith@institute.org>

The basic models are now also in place to begin ingesting data. Since your data release will likely contain catalogues, we recommend you start there first.

Read Next:

Data Types

Currently Data Central supports catalogues, IFS, spectra and imaging data. View the sections below for details on file formatting and the required metadata for each data type.

Warning

Please be sure to read the documentation carefully. Pay attention to restrictions on the contents of columns (e.g. restictions on allowed characters, choosing between a limited number of options), and double check that names match across different files. This reduces the number of issues encountered in ingestion, and reduces the number of iterations needed to finalise the data release.

Catalogues

Ingesting catalogues into Data Central allows for SQL/ADQL querying, and broadcasts the table(s) through the Data Central TAP server. The Schema Browser will also include an entry for each catalogue.

If you provide an input catalogue as part of the catalogue ingestion (as described later in this article), additional functionality is provided:

  • sources will appear in the Name Resolver and can be automatically resolved by the image cutout.

  • sources will appear in the Cone Search

  • sources will be available in the Single Object Viewer. Individual data products (IFS, Spectra) can be linked to a source, and custom SQL run to populate the SOV with particular rows from your catalogues.

Note

source in Data Central is used interchangeably with AstroObject. It is a survey-team defined astronomical object with positional information that individual data product files can be linked to.

Note

Remember that the documentation mentioned here is the static, paper-like documentation, the documentation on Documentent Central is entirely separate.

Data Model

Data Central’s ingestion process will map your data onto the Data Central data model format. Within Data Central, catalogue data are organised hierarchically, as per:

Survey
└── DataRelease
    └── Schema:Catalogues
        └── Group
            └── Table

There are dozens of tables from multiple surveys in the Data Central database. Groups are used to collect scientifically-related tables together, in order to help the user locate the correct table more quickly. To explore the data model further, visit the catalogue section of the Schema Browser to explore the relationships between groups and tables.

Directory Structure

To ingest data into Data Central, you will provide two folders, one containing the data products themselves, and one containing the metadata.

Data

The catalogues directory should contain the catalogue files themselves.

data
└── <survey>
    └── <datarelease>
        └── catalogues
            ├── my_input_cat.fits
            └── my_output_table.csv

Attention

<survey> and <datarelease> should be replaced with the values you chose in Getting Started, e.g., gama and dr2

Data Central supports catalogues/tables in .csv or .fits formats.

Danger

If your input table is > 2GB in size, please ensure the format is .csv (not fits).

Metadata

The following file structure should be adopted. A top-level <survey> directory should contain a single directory per <datarelease>. Both directories should have metadata files described below which will populate the Schema Browser.

metadata
└── <survey>
    ├── <survey>_survey_meta.txt
    └── <datarelease>
        ├── <survey>_<data_release>_data_release_meta.txt
        └── catalogues/
            ├── <survey>_<datarelease>_column_meta.txt
            ├── <survey>_<datarelease>_coordinate_meta.txt  ** optional
            ├── <survey>_<datarelease>_group_meta.txt
            ├── <survey>_<datarelease>_sql_meta.txt         ** optional
            ├── <survey>_<datarelease>_table_meta.txt
            └── docs/

The metadata/catalogues/ directory will contain a minimum of 3 metadata files, plus a docs/ directory if you have supplied additional documentation for a particular catalogue. The two optional metadata files (coordinate_meta and sql_meta) are described later in this article.

Metadata Files

Attention

Metadata files are always pipe-delimited, and have the extension .txt

<survey>_<datarelease>_group_meta.txt

This file describes the groups you would like to register, and will be used to populate the Schema Browser. Provide the following a single pipe-delimited .txt file containing an entry (row) for each group:

name

pretty_name

description

documentation

contact

date

version

ApMatchedPhotom

ApMatchedPhotom

This group provides aperture matched ugrizYJHK photometry.

unique_group_documentation_filename.txt

name <email@institute.org>

2012-04-23

v02

Please name this file: <survey>_<datarelease>_group_meta.txt e.g., sami_dr2_group_meta.txt

<survey>_<datarelease>_group_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Group name. Use only alphanumeric characters. This must be unique per data release.

pretty_name(required=True, type=char, max_limit=100)

A human-readable version of the group name. This can contain any characters (up to the character limit).

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing the group.

documentation(required=True, type=char, max_limit=1000)

If you would like formatted text to appear in the schema browser, please supply the name of the file containing html-formatted text (see Formatting for more info). Note, this is typically for 2-3 paragraphs of information. Detailed documentation should be written into a Document Central article. If you do not wish to supply documentation for a particular row, leave this entry blank.

contact(required=True, type=char, max_limit=500)

Format as: John Smith <john.smith@institute.org>

date(required=True, type=char, max_limit=100)

Group creation/update date as defined by the team e.g., 2012-04-23

version(required=True, type=char, max_limit=100)

Group version as defined by the team e.g., v1.8

<survey>_<datarelease>_table_meta.txt

This file describes the tables you would like to register, and will be used to populate the Schema Browser and be available for public SQL/ADQL querying, as well as discoverable through the Data Central TAP server.

Please provide a single .txt file with an entry per table, containing the following meta information:

name

description

documentation

group

filename

contact

date

version

ApMatchedCat

This table contains r-band aperture matched photometry and otherSource Extractor outputs for all GAMA DR2 objects.

unique_table_documentation_filename.txt

ApMatchedPhotom

ApMatchedCat.fits

name <email@institute.org>

2012-04-23

v02

Please name this file: <survey>_<datarelease>_table_meta.txt e.g., sami_dr2_table_meta.txt

<survey>_<datarelease>_table_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Table name. Use only alphanumeric characters. This must be unique per data release.

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing the group.

documentation(required=True, type=char, max_limit=1000)

If you would like formatted text to appear in the schema browser, please supply the name of the file containing html-formatted text (see Formatting for more info). Note, this is typically for 2-3 paragraphs of information. Detailed documentation should be written into a Document Central article. If you do not wish to supply documentation for a particular row, leave this entry blank.

group_name(required=True, type=char, max_limit=100)

The name of the group (must match a group name from the <survey>_<datarelease>_group_meta.txt file above)

filename(required=True, type=char, max_limit=1000)

The filename of the table you’ll be providing

contact(required=True, type=char, max_limit=500)

Format as: John Smith <john.smith@institute.org>

date(required=True, type=char, max_limit=100)

Table creation/update date as defined by the team e.g., 2012-04-23

version(required=True, type=char, max_limit=100)

Table version as defined by the team e.g., v1.8

<survey>_<datarelease>_column_meta.txt

This file describes the columns you would like to register for each table, and will be used to populate the Schema Browser, SQL/ADQL query service, and the TAP server. Please provide the following a single pipe-delimited .txt file containing an entry (row) for each column:

name

table_name

description

ucd

unit

data_type

ALPHA_J2000

ApMatchedCat

RA (r band)

pos.eq.ra;em.opt.R

deg

double

CATAID

EnvironmentMeasures

Unique GAMA ID

meta.id

double

Please name this file: <survey_dr>_column_meta.txt e.g., sami_dr2_column_meta.txt

<survey>_<datarelease>_column_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Column name. Use only alphanumeric characters.

Attention

Column names must be SQL-queriable, use only characters, letters and underscores in your column names. Column names cannot start with numbers but can include numbers afterwards. Forbidden characters include: %^&({}+-/ ][‘’’

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing the table.

table_name(required=True, type=char, max_limit=100)

The name of the table (must match a table name from the <survey>_<datarelease>_table_meta.txt file above)

ucd(required=True, type=char, max_limit=100)

UCDs can be found here: http://cds.u-strasbg.fr/UCD/tree/js/ (more info: https://arxiv.org/pdf/1110.0525.pdf)

unit(required=True, type=char, max_limit=100)

Column unit

data_type(required=True, type=char, max_limit=100)

data type of the column. Add the full name of the data type such as integer instead of shorten form int.

Extra Requirements

Note: this section is optional, you do not have to provide an _coordinate_meta.txt file or _sql_meta.txt file if your data release does not lend itself to individual source identification.

You cannot ingest individual data products associated with a single astronomical object without completing this step.

By providing a metadata file pointing to the input catalogue of your data release, Data Central is able to populate the database with Astronomical Objects from your survey. These objects are then accessible in the name resolver, and cone search (as well as the image cutout overplotting functionality).

Source Catalogue Identification

To identify a table as a source catalogue (i.e. an input catalogue that has one row per source in your data release), please provide a metadata file that contains the name of a single table that contains the resolver info (source name, coordinates, format), as per:

table_name

source_name_col

long_col

lat_col

long_format

lat_format

frame

equinox

InputCatA

CATAID

RA_deg

Dec_deg

deg

deg

icrs

Please name this file: <survey_dr>_coordinate_meta.txt e.g., sami_dr2_coordinate_meta.txt

Danger

If your input table is > 2GB in size, please ensure the format is .csv (not fits).

Tip

It is advised to provide coordinates as RA, Dec (degrees, degrees). If your Long/Lat fields are not in an ICRS coordinate frame (degrees), Data Central will auto-generate these columns.

<survey_dr>_coordinate_meta.txt

This file should contain the following columns

table_name(required=True, type=char, max_limit=100)

The table name (not filename) to be used (must have an entry in the <survey>_<datarelease>_table_meta.txt file.)

source_name_col(required=True, type=char, max_limit=100)

The column name for source name (from the specified table)

long_col(required=True, type=char, max_limit=100)

The column name for longitude (from the specified table)

lat_col(required=True, type=char, max_limit=100)

The column name for latitude (from the specified table)

long_format(required=True, type=char, max_limit=100)

The longitude format. Depending on the formatting of your coordinate values (i.e., whether decimal/space delimited/colon delimited) and the value of long_format/lat_format (deg or h), coordinate data are interpreted as:

value

format

interpretation

10.2345

deg

Degrees

1 2 3

deg

Degrees, arcmin, arcsecond

1:2:30.40

deg

Sexagesimal degrees

1 2 0

hourangle

Sexagesimal hours

lat_format(required=True, type=char, max_limit=100)

The latitude format. Depending on the formatting of your coordinate values (i.e., whether decimal/space delimited/colon delimited) and the value of long_format/lat_format (deg or h), coordinate data are interpreted as:

value

format

interpretation

10.2345

deg

Degrees

1 2 3

deg

Degrees, arcmin, arcsecond

1:2:30.40

deg

Sexagesimal degrees

1 2 0

hourangle

Sexagesimal hours

frame(required=True, type=char, max_limit=100)

Coordinate frame. Accepted values are (fk5, fk4, icrs, galactic, supergalactic)

equinox(required=True, type=char, max_limit=100)

If appropriate (leave blank for icrs), the equinox of this frame. Accepted values are (j2000, j1950, b1950)

Data Central will auto-generate ICRS-frame RA(deg) Dec(deg) columns if that format has not been provided. Data Central is able to transform using the combinations of coordinate systems listed above. If you do not see the coordinate system your data are currently recorded in, it is advised to generate RA, Dec columns as ICRS for your catalogue to be included.

Danger

The values in source_name_col must be unique across all of the tables included in the source catalogue. If none of your existing tables meet this requirement, then you will need to generate a new table, which need only include source name, RA and Dec. You do not need to include this table in the column or table metadata files, but it would be preferred.

SOV SQL Functionality

If you wish for a your tables to be queried and rows displayed as part of the Single Object Viewer, please provide the following table:

table_name

sql

InputCatA

“SELECT * FROM gama_dr2.InputCatA WHERE CATAID = {objid}”

Please name this file: <survey_dr>_sql_meta.txt e.g., sami_dr2_sql_meta.txt

<survey_dr>_sql_meta.txt

This file should contain the following columns

table_name(required=True, type=char, max_limit=100)

The table name (not filename) to be used (must have an entry in the <survey_dr>_table_meta.txt file.)

sql(required=True, type=char, max_limit=100)

The SQL expression (following the DC syntax of survey_dr.table_name) to be run on SOV load. {objid} will be replaced by the AstroObject being requested.

Attention

Ensure your SQL runs before submitting this metadata file. e.g., check whether you need single quotes around the objid as per: ‘{objid}’

IFS

Ingesting IFS data products into Data Central allows for population of the Single Object Viewer, from which users are able to download and explore (file type dependent) those data products. The Schema Browser will also include an entry for each IFS data product type. The minimum requirements for ingesting IFS data products are described below.

Note

Please Note: this is a living document, additional requirements for various data product visualizations (2D maps etc) will be set, and updated, in the near future. Although Data Central will take every effort to minimize changes to the format of the meta-data required, in order to provide survey teams with interactive visualizations and deliver the best possible user experience, we may from time-to-time require changes to these requirements.

For ingested surveys, where possible, we will auto-generate any meta data required (from fits headers), but teams should be aware that this document will likely contain additional/updated requirements on a ~6 monthly timescale in order to be compatible with the latest Data Central release.

Important

Prerequisite: You’ll need to have provided an input catalogue of sources, as per the Catalogues documentation, following all sections up to and including Source Catalogue Identification.

Note

Remember that the documentation mentioned here is the static, paper-like documentation, the documentation on Documentent Central is entirely separate.

Data Model

Data Central’s ingestion process will map your data onto the Data Central data model format. IFS data are organised hierarchically, as per:

Survey
└── DataRelease
    └── Schema:IFS
        └── Facility
            └── Data Product

To explore the data model further, visit the IFS section of the Schema Browser to explore the relationships between facilities and data products.

Directory Structure

To ingest IFS data products, you will provide two folders, one containing the data products themselves, and one containing the metadata. You’ll also need to have provided an input catalogue of sources, as per the Catalogues documentation, following all sections up to and including Source Catalogue Identification.

Data

The ifs directory should contain the fits files themselves.

data
└── <survey>
    └── <datarelease>
        └── ifs
            └── <source_name>
                ├── product1.fits
                └── product2.gz
            └── <source_name>
                ├── product3.fits
                └── product4.fits

Attention

<survey> and <datarelease> should be replaced with the values you chose in Getting Started, e.g., sami and dr2

A good rule of thumb is to keep your files succinct and with as few extensions as possible (i.e, do not pack tens of extensions into your IFS cubes, separate into different files). This makes mapping the correct file (by data type) to a browser visualizer simpler. Currently we support 2D map visualizations in the browser, have a look at a SAMI DR2 object in the Single Object Viewer.

Metadata

The following file structure should be adopted. A top-level <survey> directory should contain a single directory per <datarelease>. Both directories should have metadata files described below which will populate the Schema Browser.

metadata
└── <survey>
    ├── <survey>_survey_meta.txt
    └── <datarelease>
        ├── <survey>_<data_release>_data_release_meta.txt
        └── ifs/
            ├── <survey>_<datarelease>_facility_meta.txt
            ├── <survey>_<datarelease>_group_meta.txt
            ├── <survey>_<datarelease>_product_ingestion.txt
            ├── <survey>_<datarelease>_product_meta.txt
            └── docs/

The metadata/catalogues/ directory will contain a minimum of 4 metadata files, plus a docs/ directory if you have supplied additional documentation for a particular facility/product.

Metadata Files

Attention

Metadata files are always pipe-delimited, and have the extension .txt

<survey>_<datarelease>_facility_meta.txt

This file describes the facilities you would like to register, and will be used to populate the Schema Browser and SOV. Provide the following a single pipe-delimited .txt file containing an entry (row) for each facility:

name

pretty_name

description | documentation

sami

SAMI

Brief SAMI instrument description | sami_instrument_doc.txt

Please name this file: <survey>_<datarelease>_facility_meta.txt e.g., sami_dr2_facility_meta.txt

<survey>_<datarelease>_facility_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Facility name. Use only alphanumeric characters. This must be unique per data release.

pretty_name(required=True, type=char, max_limit=100)

A human-readable version of the facility name. This can contain any characters (up to the character limit).

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing the facility.

documentation(required=True, type=char, max_limit=1000)

If you would like formatted text to appear in the schema browser, please supply the name of the file containing html-formatted text (see Formatting for more info). Note, this is typically for 2-3 paragraphs of information. Detailed documentation should be written into a Document Central article. If you do not wish to supply documentation for a particular row, leave this entry blank.

<survey>_<datarelease>_product_meta.txt

This file describes the types of data products you would like to register, and will be used to populate the Schema Browser and SOV. Provide the following a single pipe-delimited .txt file containing an entry (row) for each product:

facility_name

name

description

documentation

group_name

version

contact

sami

spectral_cube-blue

Fully reduced and flux calibrated SAMI cube (blue)

spec_cube_blue.html

cube

1.2

John Smith <john.smith@institute.org>

sami

sfr_default_1-comp

1-component star formation rate map from default cube

sfr.html

sfr

1.2

John Smith <john.smith@institute.org>

Please name this file: <survey>_<datarelease>_product_meta.txt e.g., sami_dr2_product_meta.txt

<survey>_<datarelease>_product_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Data product name. Choose from the following:

name

pretty_name

vis_type

cube_blue

Flux Cube: Blue

3d_cube

adaptive_blue

Adaptively-binned Flux Cube: Blue

3d_cube

annular_blue

Annularly-binned Flux Cube: Blue

3d_cube

sectors_blue

Sectors-binned Flux Cube: Blue

3d_cube

cube_red

Flux Cube: Red

3d_cube

adaptive_red

Adaptively-binned Flux Cube: Red

3d_cube

annular_red

Annularly-binned Flux Cube: Red

3d_cube

sectors_red

Sectors-binned Flux Cube: Red

3d_cube

spectrum_3-kpc_blue

3kpc Aperture Spectrum: Blue

1d_spectrum

spectrum_re_blue

Re Elliptical Aperture Spectrum: Blue

1d_spectrum

spectrum_1-4-arcsec_blue

1”.4 Aperture Spectrum: Blue

1d_spectrum

spectrum_2-arcsec_blue

2” Aperture Spectrum: Blue

1d_spectrum

spectrum_3-arcsec_blue

3” Aperture Spectrum: Blue

1d_spectrum

spectrum_4-arcsec_blue

4” Aperture Spectrum: Blue

1d_spectrum

spectrum_3-kpc_red

3kpc Aperture Spectrum: Red

1d_spectrum

spectrum_re_red

Re Elliptical Aperture Spectrum: Red

1d_spectrum

spectrum_1-4-arcsec_red

1”.4 Aperture Spectrum: Red

1d_spectrum

spectrum_2-arcsec_red

2” Aperture Spectrum: Red

1d_spectrum

spectrum_3-arcsec_red

3” Aperture Spectrum: Red

1d_spectrum

spectrum_4-arcsec_red

4” Aperture Spectrum: Red

1d_spectrum

stellar-velocity_default_two-moment

Stellar Velocity Map

2d_map

stellar-velocity-dispersion_default_two-moment

Stellar Velocity Dispersion Map

2d_map

stellar-velocity_adaptive_two-moment

Stellar Velocity Map from Adaptively-binned Cube

2d_map

stellar-velocity-dispersion_adaptive_two-moment

Stellar Velocity Dispersion Map from Adaptively-binned Cube

2d_map

stellar-velocity_annular_two-moment

Stellar Velocity Map from Annularly-binned Cube

2d_map

stellar-velocity-dispersion_annular_two-moment

Stellar Velocity Dispersion Map from Annularly-binned Cube

2d_map

stellar-velocity_sectors_two-moment

Stellar Velocity Map from Sectors-binned Cube

2d_map

stellar-velocity-dispersion_sectors_two-moment

Stellar Velocity Dispersion Map from Sectors-binned Cube

2d_map

gas-velocity_default_1-comp

1-component Ionised Gas Velocity Map

2d_map

gas-vdisp_default_1-comp

1-component Ionised Gas Velocity Dispersion Map

2d_map

OII3728_default_1-comp

1-component Line Emission Map: [OII] (3726Å+3729Å)

2d_map

Halpha_default_1-comp

1-component Line Emission Map: Hα

2d_map

Hbeta_default_1-comp

1-component Line Emission Map: Hβ

2d_map

OIII5007_default_1-comp

1-component Line Emission Map: [OIII] (5007Å)

2d_map

OI6300_default_1-comp

1-component Line Emission Map: [OI] (6300Å)

2d_map

NII6583_default_1-comp

1-component Line Emission Map: [NII] (6583Å)

2d_map

SII6716_default_1-comp

1-component Line Emission Map: [SII] (6716Å)

2d_map

SII6731_default_1-comp

1-component Line Emission Map: [SII] (6731Å)

2d_map

gas-velocity_default_recom-comp

Recommended-component Ionised Gas Velocity Map

2d_map

gas-vdisp_default_recom-comp

Recommended-component Ionised Gas Velocity Dispersion Map

2d_map

OII3728_default_recom-comp

Recommended-component Line Emission Map: [OII] (3726Å+3729Å)

2d_map

Halpha_default_recom-comp

Recommended-component Line Emission Map: Hα

2d_map

Hbeta_default_recom-comp

Recommended-component Line Emission Map: Hβ

2d_map

OIII5007_default_recom-comp

Recommended-component Line Emission Map: [OIII] (5007Å)

2d_map

OI6300_default_recom-comp

Recommended-component Line Emission Map: [OI] (6300Å)

2d_map

NII6583_default_recom-comp

Recommended-component Line Emission Map: [NII] (6583Å)

2d_map

SII6716_default_recom-comp

Recommended-component Line Emission Map: [SII] (6716Å)

2d_map

SII6731_default_recom-comp

Recommended-component Line Emission Map: [SII] (6731Å)

2d_map

gas-velocity_adaptive_1-comp

1-component Ionised Gas Velocity Map from Adaptively-binned Cube

2d_map

gas-vdisp_adaptive_1-comp

1-component Ionised Gas Velocity Dispersion Map from Adaptively-binned Cube

2d_map

OII3728_adaptive_1-comp

1-component Line Emission Map from Adaptively-binned Cube: [OII] (3726Å+3729Å)

2d_map

Halpha_adaptive_1-comp

1-component Line Emission Map from Adaptively-binned Cube: Hα

2d_map

Hbeta_adaptive_1-comp

1-component Line Emission Map from Adaptively-binned Cube: Hβ

2d_map

OIII5007_adaptive_1-comp

1-component Line Emission Map from Adaptively-binned Cube: [OIII] (5007Å)

2d_map

OI6300_adaptive_1-comp

1-component Line Emission Map from Adaptively-binned Cube: [OI] (6300Å)

2d_map

NII6583_adaptive_1-comp

1-component Line Emission Map from Adaptively-binned Cube: [NII] (6583Å)

2d_map

SII6716_adaptive_1-comp

1-component Line Emission Map from Adaptively-binned Cube: [SII] (6716Å)

2d_map

SII6731_adaptive_1-comp

1-component Line Emission Map from Adaptively-binned Cube: [SII] (6731Å)

2d_map

gas-velocity_annular_2-comp

2-component Ionised Gas Velocity Map from Annularly-binned Cube

2d_map

gas-vdisp_annular_2-comp

2-component Ionised Gas Velocity Dispersion Map from Annularly-binned Cube

2d_map

OII3728_annular_2-comp

2-component Line Emission Map from Annularly-binned Cube: [OII] (3726Å+3729Å)

2d_map

Halpha_annular_2-comp

2-component Line Emission Map from Annularly-binned Cube: Hα

2d_map

Hbeta_annular_2-comp

2-component Line Emission Map from Annularly-binned Cube: Hβ

2d_map

OIII5007_annular_2-comp

2-component Line Emission Map from Annularly-binned Cube: [OIII] (5007Å)

2d_map

OI6300_annular_2-comp

2-component Line Emission Map from Annularly-binned Cube: [OI] (6300Å)

2d_map

NII6583_annular_2-comp

2-component Line Emission Map from Annularly-binned Cube: [NII] (6583Å)

2d_map

SII6716_annular_2-comp

2-component Line Emission Map from Annularly-binned Cube: [SII] (6716Å)

2d_map

SII6731_annular_2-comp

2-component Line Emission Map from Annularly-binned Cube: [SII] (6731Å)

2d_map

gas-velocity_sectors_1-comp

1-component Ionised Gas Velocity Map from Sectors-binned Cube

2d_map

gas-vdisp_sectors_1-comp

1-component Ionised Gas Velocity Dispersion Map from Sectors-binned Cube

2d_map

OII3728_sectors_1-comp

1-component Line Emission Map from Sectors-binned Cube: [OII] (3726Å+3729Å)

2d_map

Halpha_sectors_1-comp

1-component Line Emission Map from Sectors-binned Cube: Hα

2d_map

Hbeta_sectors_1-comp

1-component Line Emission Map from Sectors-binned Cube: Hβ

2d_map

OIII5007_sectors_1-comp

1-component Line Emission Map from Sectors-binned Cube: [OIII] (5007Å)

2d_map

OI6300_sectors_1-comp

1-component Line Emission Map from Sectors-binned Cube: [OI] (6300Å)

2d_map

NII6583_sectors_1-comp

1-component Line Emission Map from Sectors-binned Cube: [NII] (6583Å)

2d_map

SII6716_sectors_1-comp

1-component Line Emission Map from Sectors-binned Cube: [SII] (6716Å)

2d_map

SII6731_sectors_1-comp

1-component Line Emission Map from Sectors-binned Cube: [SII] (6731Å)

2d_map

gas-velocity_adaptive_recom-comp

Recommended-component Ionised Gas Velocity Map from Adaptively-binned Cube

2d_map

gas-vdisp_adaptive_recom-comp

Recommended-component Ionised Gas Velocity Dispersion Map from Adaptively-binned Cube

2d_map

OII3728_adaptive_recom-comp

Recommended-component Line Emission Map from Adaptively-binned Cube: [OII] (3726Å+3729Å)

2d_map

Halpha_adaptive_recom-comp

Recommended-component Line Emission Map from Adaptively-binned Cube: Hα

2d_map

Hbeta_adaptive_recom-comp

Recommended-component Line Emission Map from Adaptively-binned Cube: Hβ

2d_map

OIII5007_adaptive_recom-comp

Recommended-component Line Emission Map from Adaptively-binned Cube: [OIII] (5007Å)

2d_map

OI6300_adaptive_recom-comp

Recommended-component Line Emission Map from Adaptively-binned Cube: [OI] (6300Å)

2d_map

NII6583_adaptive_recom-comp

Recommended-component Line Emission Map from Adaptively-binned Cube: [NII] (6583Å)

2d_map

SII6716_adaptive_recom-comp

Recommended-component Line Emission Map from Adaptively-binned Cube: [SII] (6716Å)

2d_map

SII6731_adaptive_recom-comp

Recommended-component Line Emission Map from Adaptively-binned Cube: [SII] (6731Å)

2d_map

extinct-corr_default_1-comp

Extinction correction map from 1-component Hα/Hβ flux ratio

2d_map

sfr-mask_default_1-comp

Star formation mask map from 1-component emission line ratios

2d_map

sfr_default_1-comp

Star formation rate map from 1-component Hα flux

2d_map

sfr-dens_default_1-comp

Star formation rate surface density from 1-component Hα flux

2d_map

extinct-corr_default_recom-comp

Extinction correction map from recommended-component Hα/Hβ flux ratio

2d_map

sfr-mask_default_recom-comp

Star formation mask map from recommended-component emission line ratios

2d_map

sfr_default_recom-comp

Star formation rate map from recommended-component Hα flux

2d_map

sfr-dens_default_recom-comp

Star formation rate surface density from recommended-component Hα flux

2d_map

gas-velocity_sectors_recom-comp

Recommended-component Ionised Gas Velocity Map from Sectors-binned Cube

2d_map

gas-vdisp_sectors_recom-comp

Recommended-component Ionised Gas Velocity Dispersion Map from Sectors-binned Cube

2d_map

OII3728_sectors_recom-comp

Recommended-component Line Emission Map from Sectors-binned Cube: [OII] (3726Å+3729Å)

2d_map

Halpha_sectors_recom-comp

Recommended-component Line Emission Map from Sectors-binned Cube: Hα

2d_map

Hbeta_sectors_recom-comp

Recommended-component Line Emission Map from Sectors-binned Cube: Hβ

2d_map

OIII5007_sectors_recom-comp

Recommended-component Line Emission Map from Sectors-binned Cube: [OIII] (5007Å)

2d_map

OI6300_sectors_recom-comp

Recommended-component Line Emission Map from Sectors-binned Cube: [OI] (6300Å)

2d_map

NII6583_sectors_recom-comp

Recommended-component Line Emission Map from Sectors-binned Cube: [NII] (6583Å)

2d_map

SII6716_sectors_recom-comp

Recommended-component Line Emission Map from Sectors-binned Cube: [SII] (6716Å)

2d_map

SII6731_sectors_recom-comp

Recommended-component Line Emission Map from Sectors-binned Cube: [SII] (6731Å)

2d_map

extinct-corr_adaptive_1-comp

Extinction correction map from 1-component adaptive-binned Hα/Hβ flux ratio

2d_map

sfr-mask_adaptive_1-comp

Star formation mask map from 1-component adaptive-binned emission line ratios

2d_map

sfr_adaptive_1-comp

Star formation rate map from 1-component adaptive-binned Hα flux

2d_map

sfr-dens_adaptive_1-comp

Star formation rate surface density from 1-component adaptive-binned Hα flux

2d_map

extinct-corr_adaptive_recom-comp

Extinction correction map from recommended-component adaptive-binned Hα/Hβ flux ratio

2d_map

sfr-mask_adaptive_recom-comp

Star formation mask map from recommended-component adaptive-binned emission line ratios

2d_map

sfr_adaptive_recom-comp

Star formation rate map from recommended-component adaptive-binned Hα flux

2d_map

sfr-dens_adaptive_recom-comp

Star formation rate surface density from recommended-component adaptive-binned Hα flux

2d_map

extinct-corr_sectors_1-comp

Extinction correction map from 1-component sectors-binned Hα/Hβ flux ratio

2d_map

sfr-mask_sectors_1-comp

Star formation mask map from 1-component sectors-binned emission line ratios

2d_map

sfr_sectors_1-comp

Star formation rate map from 1-component sectors-binned Hα flux

2d_map

sfr-dens_sectors_1-comp

Star formation rate surface density from 1-component sectors-binned Hα flux

2d_map

extinct-corr_sectors_recom-comp

Extinction correction map from recommended-component sectors-binned Hα/Hβ flux ratio

2d_map

sfr-mask_sectors_recom-comp

Star formation mask map from recommended-component sectors-binned emission line ratios

2d_map

sfr_sectors_recom-comp

Star formation rate map from recommended-component sectors-binned Hα flux

2d_map

sfr-dens_sectors_recom-comp

Star formation rate surface density from recommended-component sectors-binned Hα flux

2d_map

extinct-corr_annular_2-comp

Extinction correction map from 2-component annular-binned Hα/Hβ flux ratio

2d_map

sfr-mask_annular_2-comp

Star formation mask map from 2-component annular-binned emission line ratios

2d_map

sfr_annular_2-comp

Star formation rate map from 2-component annular-binned Hα flux

2d_map

sfr-dens_annular_2-comp

Star formation rate surface density from 2-component annular-binned Hα flux

2d_map

If you have additional product names that aren’t covered by this list, please let us know as soon as possible.

facility_name(required=True, type=char, max_limit=100)

The name of the facility (must match a facility name from the <survey>_<datarelease>_facility_meta.txt file)

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing the product.

documentation(required=True, type=char, max_limit=1000)

If you would like formatted text to appear in the schema browser, please supply the name of the file containing html-formatted text (see Formatting for more info). Note, this is typically for 2-3 paragraphs of information. Detailed documentation should be written into a Document Central article. If you do not wish to supply documentation for a particular row, leave this entry blank.

group_name(required=True, type=char, max_limit=100)

The name of the group (must match a group name from the <survey>_<datarelease>_group_meta.txt file)

version(required=True, type=char, max_limit=100)

Product version as defined by the team e.g., v1.8

contact(required=True, type=char, max_limit=500)

Format as: John Smith <john.smith@institute.org>

<survey>_<datarelease>_group_meta.txt

To aid the user in finding data products, a grouping is applied in the Schema Browser and the SOV (i.e., all IFS-derived spectra can be grouped under ‘Spectra’). This file describes the groups you would like to register, and will be used to populate the Schema Browser and categorize data products in the SOV. Provide the following a single pipe-delimited .txt file containing an entry (row) for each group:

name

pretty_name

cube

Cubes

stelkin

Stellar Kinematics

Please name this file: <survey>_<datarelease>_group_meta.txt e.g., sami_dr2_group_meta.txt

<survey>_<datarelease>_group_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Group name. Use only alphanumeric characters. This must be unique per data release.

pretty_name(required=True, type=char, max_limit=100)

A human-readable version of the group name. This can contain any characters (up to the character limit).

<survey>_<datarelease>_product_ingestion.txt

The final file you will need to provide links each data product with the relevant source from your input catalogue (as mentioned above, you’ll need to have provided an input catalogue of sources, as per the Catalogues documentation, following all sections up to and including Source Catalogue Identification). This file should include an entry (row) for every file you wish to ingest into Data Central.

facility_name

data_product_name

file_name

rel_file_path

source_name

sami

OIII5007_annular_2-comp

99428_OIII5007_annular_2-comp.fits

99428/99428_OIII5007_annular_2-comp.fits

99428

Please name this file: <survey_dr>_product_ingestion.txt e.g., sami_dr2_product_ingestion.txt

<survey>_<datarelease>_product_ingestion.txt

This file should contain the following columns

facility_name(required=True, type=char, max_limit=100)

The name of the facility (must match a facility name from the <survey>_<datarelease>_facility_meta.txt file)

data_product_name(required=True, type=char, max_limit=100)

The name of the data product (must match a facility name from the <survey>_<datarelease>_product_meta.txt file)

file_name(required=True, type=char, max_limit=100)

The filename of the data product you’ll be providing

rel_file_path(required=True, type=char, max_limit=100)

The relative path of the file. e.g., <source_name>/product1.fits. See Directory structure.

source_name(required=True, type=char, max_limit=100)

The source name as provided in your source catalogue (see Source Catalogue Identification).

is_best(required=False, type=bool, default=True)

You may associate multiple IFS with a single product, and define ONE of those products as the best (add a description of this in your <survey>_<datarelease>_product_meta.txt file or data release information on Document Central. The best product for a source will be highlighted in the SOV as such.

Images

Ingesting fits images allows for their retrieval via the a multi-band image cutout service. This service allows astronomers to easily and intuitively access imaging data from some of the world’s leading facilities alongside datasets of national significance, to quickly explore sources of interest and produce publication quality images.

The image cutout service allows for comparison of the same source at different wavelengths side by side. Users can input a single source/position or a list (automatically transforming between the submitted coordinate frame and the native coordinate frame of the image) to generate single bands and/or three-colour images.

Five image scaling options are supported, along with custom plot options (axes style, colour map, coordinate grid), over-plotting source positions from any of the DC-hosted input catalogues as well as remote VO datasets, and over-plotting contours for a particular band at user-specified levels. Users are able to download results as .png or .fits files, on a per-band/per-source basis, or download all the files from one request.

The cutout can host mosaiced image files in a range of projections and pixel scales, automatically reprojecting data on the fly.

Note

Remember that the documentation mentioned here is the static, paper-like documentation, the documentation on Documentent Central is entirely separate.

Data Model

Data Central’s ingestion process will map your data onto the Data Central data model format. Within Data Central, imaging files are organised hierarchically, as per:

Survey
└── DataRelease
    └── Schema:Imaging
        └── Facility
            └── Band
                └── Region

Currently, the cutout service is not able to stitch together imaging on the fly. You must provide FITS images that have been resampled and co-added (see for example, SWarp.) Each file is saved as a Region on the sky, with the fits header stored in the database to provide the coordinate lookup. Note, this means individual region files must uniquely cover one area of sky. Data central can not currently combine or display multiple images in the same band covering the sam sky region. To explore the data model further, visit the imaging section of the Schema Browser to explore the relationships between facilities and bands.

Directory Structure

To ingest data into Data Central, you will provide two folders, one containing the imaging files themselves, and one containing the metadata.

Data

The imaging directory should contain the fits files themselves.

data
└── <survey>
    └── <datarelease>
        └── imaging
            ├── swpim_galex_FUV_d10.fits
            └── swpim_galex_NUV_g09.fits

Attention

<survey> and <datarelease> should be replaced with the values you chose in Getting Started, e.g., gama and dr2

Data Central supports images in .fits formats. Note, these data files can technically have any file name, as long as they are correctly matching in your metadata files (see below). However, we strongly suggest using sensible file names that describe the facility, band, survey and region.

Metadata

The following file structure should be adopted. A top-level <survey> directory should contain a single directory per <datarelease>. Both directories should have metadata files described below which will populate the Schema Browser.

metadata
└── <survey>
    ├── <survey>_survey_meta.txt
    └── <datarelease>
        ├── <survey>_<data_release>_data_release_meta.txt
        └── imaging/
            ├── <survey>_<datarelease>_facility_meta.txt
            ├── <survey>_<datarelease>_band_meta.txt
            └── docs/

The metadata/imaging/ directory will contain 2 metadata files, plus a docs/ directory if you have supplied additional documentation for a particular facility/band.

Metadata Files

Attention

Metadata files are always pipe-delimited, and have the extension .txt

<survey>_<datarelease>_facility_meta.txt

This file describes the facilities you would like to register, and will be used to populate the Schema Browser. Provide the following a single pipe-delimited .txt file containing an entry (row) for each facility:

name

pretty_name

description

documentation

CFHTCFHTCOSMOS

CFHT

Images from the CFHT COSMOS survey.

unique_cfht_documentation_filename.txt

Please name this file: <survey>_<datarelease>_facility_meta.txt e.g., devils_dr1_facility_meta.txt

<survey>_<datarelease>_facility_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Facility name. Use only alphanumeric characters. This must be unique per data release.

pretty_name(required=True, type=char, max_limit=100)

A human-readable version of the facility name. This can contain any characters (up to the character limit).

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing the facility.

documentation(required=True, type=char, max_limit=1000)

If you would like formatted text to appear in the schema browser, please supply the name of the file containing html-formatted text (see Formatting for more info). Note, this is typically for 2-3 paragraphs of information. Detailed documentation should be written into a Document Central article. If you do not wish to supply documentation for a particular row, leave this entry blank.

<survey>_<datarelease>_band_meta.txt

This file describes the bands you would like to register, and will be used to populate the Schema Browser and be available to select in the cutout service.

Please provide a single .txt file with an entry per file (each file should have a corresponding unique ‘region’), containing the following meta information:

name

pretty_name

start_wave_ang

central_wave_ang

end_wave_ang

spatial_res

start_time

end_time

pol_state

facility

description

documentation

contact

version

file_name

region

is_displayed

H

H

14629

16499.11

18085

0.5

CFHTCFHTCOSMOS

COSMOS H-Band

unique_h-band_doc.txt

name <email@institute.org>

v02

swpim_cfht_H_d10.fits

d10

Y

Please name this file: <survey>_<datarelease>_band_meta.txt e.g., devils_dr1_band_meta.txt

<survey>_<datarelease>_band_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Band name. This must be unique per data release.

pretty_name(required=True, type=char, max_limit=100)

A human-readable version of the band name. This can contain any characters (up to the character limit).

start_wave_ang(required=True, type=char, max_limit=100)

A wavelength for the start of the filter bandpass of the image - in Angstrom. This is required by the Simple Image Access service.

central_wave_ang(required=True, type=char, max_limit=100)

A rough central wavelength of the image - in Angstrom. This is used to order the bands in the cutout service options.

end_wave_ang(required=True, type=char, max_limit=100)

A wavelength for the end of the filter bandpass of the image - in Angstrom. This is required by the Simple Image Access service.

spatial_res(required=True, type=char, max_limit=100)

The spatial resolution (FWHM) of the image - in arcsec. This is required by the Simple Image Access service.

start_time(required=True, type=char, max_limit=100)

The start time of the image observation in UTC isoformat: yyyy-mm-ddThh:mm:ss.ss. In case of mosaics made of multiple images, this may either be left blank or preferably correspond to the start time of the earliest image in the mosaic. This is required by the Simple Image Access service.

end_time(required=True, type=char, max_limit=100)

The end time of the image observation in UTC isoformat: yyyy:mm:ddThh:mm:ss.ss. In case of mosaics made of multiple images, this may either be left blank or preferably correspond to the end time of the latest image in the mosaic. This is required by the Simple Image Access service.

pol_state(required=True, type=char, max_limit=100)

The polarization state(s) of the image. An empty string if irrelevant (e.g. optical data), otherwise a space-separated string consisting of one or more of the following: I Q U V RR LL RL LR XX YY XY YX POLI POLA. This is required by the Simple Image Access service.

facility(required=True, type=char, max_limit=100)

The name of the facility (must match a facility name from the <survey>_<datarelease>_facility_meta.txt file above)

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing the facility.

documentation(required=True, type=char, max_limit=1000)

If you would like formatted text to appear in the schema browser, please supply the name of the file containing html-formatted text (see Formatting for more info). Note, this is typically for 2-3 paragraphs of information. Detailed documentation should be written into a Document Central article. If you do not wish to supply documentation for a particular row, leave this entry blank.

contact(required=True, type=char, max_limit=500)

Format as: John Smith <john.smith@institute.org>

version(required=True, type=char, max_limit=100)

Band version as defined by the team e.g., v1.8

file_name(required=True, type=char, max_limit=1000)

The filename of the fits file you’ll be providing. Note, you can provide multiple files at the same wavelength by providing an entry per file.

region(required=True, type=char, max_limit=1000)

The name of the region for this file (e.g., g09, g12). You may choose to define this from the center RA of the image.

is_displayed(required=True, type=bool)

Y or N. If N, this row will be skipped on ingestion (you may wish to give us the meta data now and the file later).

Attention

The fits header must contain the following keywords:

Keyword

Description

GAIN

Maximum equivalent gain (e-/ADU)

CD1_1

Linear projection matrix

CD1_2

Linear projection matrix

CD2_1

Linear projection matrix

CD2_2

Linear projection matrix

NAXIS

Number of axes

CRPIX1

Reference pixel on this axis

CRPIX2

Reference pixel on this axis

CRVAL1

World coordinate on this axis

CRVAL2

World coordinate on this axis

CTYPE1

WCS projection type for this axis

CTYPE2

WCS projection type for this axis

CUNIT1

Axis unit

CUNIT2

Axis unit

NAXIS1

Number of elements along this axis

NAXIS2

Number of elements along this axis

EQUINOX

Mean equinox

RADECSYS

Astrometric system

Spectra

Ingesting spectral data products into Data Central allows for population of the Single Object Viewer, and enables download of those data product files. The Schema Browser will also include an entry for each spectral data product type. The minimum requirements for ingesting spectral data products are described below.

Note

Please Note: this is a living document, additional requirements for various data product visualizations (1D spectra etc) will be set, and updated, in the near future. Although Data Central will take every effort to minimize changes to the format of the meta-data required, in order to provide survey teams with interactive visualizations and deliver the best possible user experience, we may from time-to-time require changes to these requirements.

For ingested surveys, where possible, we will auto-generate any meta data required (from fits headers), but teams should be aware that this document will likely contain additional/updated requirements on a ~6 monthly timescale in order to be compatible with the latest Data Central release.

Important

Prerequisite: You’ll need to have provided an input catalogue of sources, as per the Catalogues documentation, following all sections up to and including Source Catalogue Identification.

Note

Remember that the documentation mentioned here is the static, paper-like documentation, the documentation on Documentent Central is entirely separate.

Data Model

Data Central’s ingestion process will map your data onto the Data Central data model format. Spectra are organised hierarchically, as per:

Survey
└── DataRelease
    └── Schema:Spectra
        └── Facility
            └── Data Product

To explore the data model further, visit a survey with spectra (e.g., GAMA DR2) in the Schema Browser to explore the relationships between facilities and data products.

Directory Structure

To ingest spectra, you will provide two folders, one containing the data products themselves, and one containing the metadata. You’ll also need to have provided an input catalogue of sources, as per the Catalogues documentation, following all sections up to and including Source Catalogue Identification.

Data

The spectra directory should contain the fits files themselves.

data
└── <survey>
    └── <datarelease>
        └── spectra
            └── <facility_name>
                ├── product1.fits
                └── product2.gz
            └── <facility_name>
                ├── product3.fits
                └── product4.fits

Attention

<survey> and <datarelease> should be replaced with the values you chose in Getting Started, e.g., gama and dr2

Data Central supports spectra .fits format only (single file per spectrum).

A good rule of thumb is to keep your files succinct and with as few extensions as possible (i.e,, do not pack tens of extensions into your fits files). This makes mapping the correct file (by data type) to a browser visualizer simpler.

Metadata

The following file structure should be adopted. A top-level <survey> directory should contain a single directory per <datarelease>. Both directories should have metadata files described below which will populate the Schema Browser.

metadata
└── <survey>
    ├── <survey>_survey_meta.txt
    └── <datarelease>
        ├── <survey>_<data_release>_data_release_meta.txt
        └── spectra/
            ├── <survey>_<datarelease>_facility_meta.txt
            ├── <survey>_<datarelease>_product_ingestion.txt
            ├── <survey>_<datarelease>_product_meta.txt
            └── docs/

The metadata/catalogues/ directory will contain a minimum of 3 metadata files, plus a docs/ directory if you have supplied additional documentation for a particular facility/product.

Metadata Files

Attention

Metadata files are always pipe-delimited, and have the extension .txt

<survey>_<datarelease>_facility_meta.txt

Provide the following a single pipe-delimited txt file containing an entry (row) for each facility:

name

pretty_name

description | documentation

2dfgrs

2dFGRS

All 2dFGRS spectra of GAMA DR2 objects obtained through the 2dFGRS query page. | 2dFGRS_survey.html

Please name this file: <survey>_<datarelease>_facility_meta.txt e.g., gama_dr2_facility_meta.txt

<survey>_<datarelease>_facility_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Facility name. Use only alphanumeric characters. This must be unique per data release.

pretty_name(required=True, type=char, max_limit=100)

A human-readable version of the facility name. This can contain any characters (up to the character limit).

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing the facility.

documentation(required=True, type=char, max_limit=1000)

If you would like formatted text to appear in the schema browser, please supply the name of the file containing html-formatted text (see Formatting for more info). Note, this is typically for 2-3 paragraphs of information. Detailed documentation should be written into a Document Central article. If you do not wish to supply documentation for a particular row, leave this entry blank.

<survey>_<datarelease>_product_meta.txt

All types of data products (regardless of whether a particular astro object has the data product) will appear in the Schema Browser. Provide the following a single pipe-delimited txt file containing an entry (row) for each product:

facility_name

name

description

documentation

version

contact

2dfgrs

spectrum_1d

Reduced 1D spectrum

2dfgrs_spec.html

1.2

John Smith <john.smith@institute.org>

2qz

spectrum_1d

Reduced 1D spectrum

1.2

John Smith <john.smith@institute.org>

Please name this file: <survey_dr>_product_meta.txt e.g., gama_dr2_product_meta.txt

Attention

Depending on the value for the name column, you will need to add additional columns to this table. See File Formats for more information.

<survey>_<datarelease>_product_meta.txt

This file should contain the following columns

name(required=True, type=char, max_limit=100)

Data product name. Choose from the following:

name

value

vis_type

spectrum_1d

1D Spectrum

1d_spectrum

spectrum_2d

2D Spectrum

2d_spectrum

Now check the File Formats section for any additional columns you may need to add to this file. If you have additional product names that aren’t covered by this list, please let us know as soon as possible.

facility(required=True, type=char, max_limit=100)

The name of the facility (must match a facility name from the <survey>_<datarelease>_facility_meta.txt file)

description(required=True, type=char, max_limit=1000)

A succinct paragraph describing the product.

documentation(required=True, type=char, max_limit=1000)

If you would like formatted text to appear in the schema browser, please supply the name of the file containing html-formatted text (see Formatting for more info). Note, this is typically for 2-3 paragraphs of information. Detailed documentation should be written into a Document Central article. If you do not wish to supply documentation for a particular row, leave this entry blank.

version(required=True, type=char, max_limit=100)

Product version as defined by the team e.g., v1.8

contact(required=True, type=char, max_limit=500)

Format as: John Smith <john.smith@institute.org>

data_format(required=True, type=char, max_limit=100)

This describes the format of the data, and tells Data Central how to read the files from the facility. Note that all the files from a facility must be in the same format (you can create multiple facilities if needed).

This can be one of three options:

  • Single-Split: the data is stored in multiple HDUs, with each HDU composed of a single 1D array.

  • Multiline-Single: the data is stored in a single HDU, across multiple lines (i.e. a 2D array).

  • specutils: the data is in neither of the above formats, and should be loaded via a specific specutils_ loader. If you think this might be the case, please contact us as soon as possible, as this option is somewhat more complex than the first two.

If specutils is selected, then add additional column specutils_format, which is the name of the specutils_ loader to use to read the file.

If Multiline-Single or Single-Split are selected, then add the additional columns all_keywords, all_standard_units, valid_wcs and (if needed, see below) fixup_file.

specutils_format(required=True, type=char, max_limit=100)

This is the name of the specutils loader to use to read the fits file. A list of possible loaders can be found here, and instructions for creating new loaders can be found here.

valid_wcs(required=True, type=bool)

Is the WCS valid (e.g. do the axes match what is expected, do they meet the WCS standard)? Either true or false. The best strategy to check this is to use astropy to try to read the WCS from the FITS header.

If false, fixup_file must be set.

all_standard_units(required=True, type=bool)

Are all the required units in the header, with the right keyword? Either true or false. In general, this means that both CUNIT1 (for the spectral unit, e.g. wavelength or frequency) and BUNIT (for the flux unit) are set. Try reading these header keywords with astropy if you are unsure.

If false, fixup_file must be set.

all_keywords(required=True, type=bool)

Do the HDU(s) include the required keywords to determine their purpose? Either true or false.

We look at different keywords depending on Single-Split or Multiline-Single.

For Multiline-Single, either the EXTNAME or HDUNAME keyword is used, whereas for Single-Split, either the ROW or ARRAY keywords are used.

Using the keyword, the value of the keyword is matched to those in the table below (in a case-insensitive manner).

Value

Purpose

badpix

Skip this row/HDU, do not treat as a spectra.

blank value

Skip this row/HDU, do not treat as a spectra.

sky

Treat as sky spectra, matching to last science spectra read.

stdev

Treat as standard deviation on the last science spectra read.

sigma

Treat as standard deviation on the last science spectra read.

variance

Treat as variance on the last science spectra read.

spectrum

Treat as a science spectra.

These keywords are then mapped to one of the purposes below (the following can also be used as the value in the FITS file).

Name

Purpose

skip

Skip this HDU/row, don’t try to interprate it as a spectrum (e.g. is an image or table).

science

The actual science spectrum.

error_*

The error on a science spectrum. It is assumed that this applies to the previous science spectrum. If a different scheme is needed, a specutils loader should be used. Currently the possible values are error-stdev, error-variance and error-inversevariance, which come from what error handling is supported by astropy. Other error schemes should add support to astropy, or mark the row/HDU as skip.

combined_*

A science or error spectrum combined from other spectra within the file. This will be used preferentially over other spectra found, so should only appear once.

sky

Spectra of the sky (presumably used for reduction purposes).

unreduced_*

Unreduced science or error, will preferentially not be used over other spectra, unless all spectra are unreduced (unlikely).

If false (or you would like to specify more detail about the purpose of each HDU or row), fixup_file must be set.

fixup_file(required=True, type=char, max_limit=100)

The fixup file is a YAML file which includes additional metadata about how to read and interpret your spectra, and the format is documented below. fixup_file is the relative path to this file. You can have one fixup file per facility.

The fixup file

The fixup file is a YAML file with a specific set of information which tells Data Central how to understand your spectra. YAML is common human readable and writeable configuration format (vs. JSON or similar which while readable are quite hard to write).

If you have not seen or written YAML before, https://camel.readthedocs.io/en/latest/yamlref.html is a good reference—we only use a small subset of the features of YAML, so don’t feel you need to understand everything on that page.

There are three sections in the fixup file, wcs, units and hdus/hdu.

wcs sets which FITS header keywords or the values to use for WCS. The following keys are used within the fixup file (Note that either the value or keyword should be specified. Otherwise all keys are required.):

  • pixel_reference_point_keyword or pixel_reference_point: The reference pixel to use, this is usually set by the FITS keyword CRPIX1.

  • pixel_reference_point_value_keyword or pixel_reference_point_value: The value of the reference pixel, this is usually set by the FITS keyword CRVAL1.

  • pixel_width_keyword or pixel_width: The width of each pixel in the WCS unit, this is usually set by the FITS keyword CDELT1.

  • wavelength_unit_keyword or wavelength_unit: The units that the WCS values are in, this is usually set by the FITS keyword CUNIT1.

units sets which FITS header keywords or the values to use for the flux. The following keys are used within the fixup file (Note that either the value or keyword should be specified.):

  • flux_unit_keyword or flux_unit: The units the spectral flux is in. This is usually set by the BUNIT FITS keyword. One of these keywords is required.

  • flux_scale_keyword or flux_scale: How much to scale the flux by. This is usually set by the BSCALE FITS keyword, and if not specified is treated as 1.

The third and most complex section is the hdus/hdu section. hdus is used by the Single-Split format, whereas hdu is used by the Multiline-Single format. Both systems allow for quite precise specification of how the data should be treated. Ideally, this section isn’t needed, and your FITS headers contain sufficient metadata that everything can be automatically detected. Also, this section can only do so much—if there needs to be different metadata specified for each individual FITS file, then this section is insufficient. In that case, creating a specutils loader might be the best way forward.

For the hdu section, you can specify some metadata for the whole HDU, and some additional data on a per-row basis. For the whole hdu, the following options are available:

  • require_transpose: Defaults to False, this is if the data is stored in columns rather than rows.

  • purpose_prefix: Which keywords hold the purpose of each row. Should be of the form <PREFIX>n e.g. ROW0 ROW1 etc. Must cover all rows. The valid values can be found under the all_keywords column.

For each row, in addition to the keywords under the wcs and units sections, the purpose key can be set, and follows the same rules as all_keywords.

The hdus sections is very similar to the hdu section, though rather than dealing with rows, it deals with HDUs. For the whole file, a purpose_prefix can be set (following the same rules given under the hdu section), and for each HDU a purpose key can be set (again using the same rules as the hdu section).

Finally, under the hdu or hdus section, a single cycle key can be set. This follows the same rules as a normal hdu or hdus section (other than it cannot contain another cycle), and allows specifying metadata when there may be a variable number of rows/HDUs with a single collective purpose. An example fixup files (from the OzDES survey) is below:

hdus:
  "0":
    purpose: "combined_science"
  "1":
    purpose: "combined_error_variance"
  "2":
    purpose: "skip"
  "cycle":
    "0":
      purpose: "science"
    "1":
      purpose: "error_variance"
    "2":
      purpose: "skip"
<survey>_<datarelease>_product_ingestion.txt

The final file you will need to provide links each data product with the relevant source from your input catalogue (as mentioned above, you’ll need to have provided an input catalogue of sources, as per the Catalogues documentation, following all sections up to and including Source Catalogue Identification). This file should include an entry (row) for every file you wish to ingest into Data Central.

facility_name

data_product_name

file_name

rel_file_path

source_name

specid

is_best

hdu

purpose

ra

dec

wmin

wmax

start_time

end_time

z

hrv

target_name

snr

resolving_power

spatial_res

gama

spectrum_1d

G12_Y3_017_187.fit

gama/G12_Y3_017_187.fit

6802

G12_Y3_017_187

TRUE

0

science

174.006

0.72093

3727.71

8857.67

2009-01-01T00:00:00.00Z

2009-01-01T00:30:00.00Z

0.05

GAMAJ113601.43+004315.3

3.37

1000

2.1

Please name this file: <survey_dr>_product_ingestion.txt e.g., gama_dr2_product_ingestion.txt

<survey>_<datarelease>_product_ingestion.txt

This file should contain the following columns

facility_name(required=True, type=char, max_limit=100)

The name of the facility (must match a facility name from the <survey>_<datarelease>_facility_meta.txt file)

data_product_name(required=True, type=char, max_limit=100)

The name of the data product (must match a facility name from the <survey>_<datarelease>_product_meta.txt file)

file_name(required=True, type=char, max_limit=100)

The filename of the data product you’ll be providing

rel_file_path(required=True, type=char, max_limit=100)

The relative path of the file. e.g., <facility_name>/product1.fits. See Directory structure.

source_name(required=True, type=char, max_limit=100)

The source name as provided in your source catalogue (see Source Catalogue Identification).

specid(required=False, type=char, max_limit=60)

A name/identifier used to distinguish between different spectra with the same source_name. Can include the source name, but that is not required.

is_best(required=False, type=bool, default=True)

You may associate multiple spectra with a single source, and define ONE of those spectra as the best (add a description of this in your <survey>_<datarelease>_product_meta.txt file or data release information on Document Central. The best spectra for a source will be highlighted in the SOV as such.

ra(required=True, type=char, max_limit=30)

The RA of the spectrum (icrs) in decimal degrees. This is required by the Simple Spectrum Access service.

dec(required=True, type=char, max_limit=30)

The Declination of the spectrum (icrs) in decimal degrees. This is required by the Simple Spectrum Access service.

Simple Spectral Access Metadata

In addition to the other metadata needed to load and read the spectra files, surveys need to provide metadata such that the spectra files can be served over the Simple Spectral Access Protocol (SSA). This metadata can either be in the individual FITS files (preferred), or provided by additional files as part of the metadata preparation process.

<survey>_<datarelease>_product_ingestion.txt
start_time(required=True, type=char, max_limit=50)

The start time of the observation in UTC isoformat: yyyy-mm-ddThh:mm:ss.ss. If the spectrum is the combination of multiple spectra, use the start time of the earliest spectrum. By default, the FITS keyword DATE-OBS will be looked at.

end_time(required=True, type=char, max_limit=50)

The end time of the observation in UTC isoformat: yyyy-mm-ddThh:mm:ss.ss. If the spectrum is the combination of multiple spectra, use the end time of the latest spectrum. By default, the FITS keyword DATE-END will be looked at.

wmin(required=True, type=float)

The start wavelength of the spectrum - in Angstrom. This should be computable via the WCS information within the file.

wmax(required=True, type=float)

The end wavelength of the spectrum - in Angstrom. This should be computable via the WCS information within the file.

z(required=False, type=float)

The determined redshift of the spectrum. By default, the FITS keyword Z will be looked at.

hrv(required=False, type=float)

The determined heliocentric radial velocity of the spectrum - in km/s. By default, the FITS keyword TODO will be looked at. TODO

target_name(required=False, type=char, max_limit=150)

The preferred target name of the spectrum as a string. By default, the FITS keyword OBJECT will be looked at.

snr(required=False, type=float)

A representative signal-to-noise ratio of the spectrum (e.g. median or at the central wavelength). By default, the FITS keyword TODO will be looked at. TODO

resolving_power(required=False, type=float)

The resolving power (lambda/delta lambda) of the spectrum at its central wavelength. By default, the FITS keyword TODO will be looked at. TODO

spatial_res(required=False, type=float)

The spatial resolution corresponding to the PSF of the observed spectrum - in arcsec. By default, the FITS keyword TODO will be looked at. TODO

If a different FITS keyword should be read, an additional section ssa can be added to the approriate fixup file (in the case of Multiline-Single or Single-Split format) or by adding an ssa dictionary to the meta object in the specutils loader.

Data File Formats

All FITS files

FITS files can themselves contain significant amounts of metadata, via the header in each extension. Using the correct headers and including all the relevant metadata both makes it easier for Data Central to ingest and manage your survey data, and reduce the amount of issues that users of the survey will encounter. To this end, here are some things you can do to ensure that the FITS files are in the best condition:

  1. astropy contains tools to ensure correctness of FITS files and metadata, run these tools on your FITS files before uploading them to Data Central. The ones we suggest running at a minimum are wcslist (for WCS) and verify() (for the FITS file itself).

  2. Sanity check that the WCS metadata contained within the file: there may be keywords left over from the data reduction process (e.g. a 2d WCS for a 1d spectrum), or tools adding additional WCS keywords which override earlier values (e.g. CRPIXjs vs. CDi_js vs. PCi_js). astropy may be able to pick up some of these issues, but depending on how the data is stored, this is not always possible (e.g. a 2d WCS would not be detected for spectra stored in rows within a single HDU, whilst logically each row should be treated as being independent). A common cause of this is copying the contents of the FITS header from unreduced to a reduced version of the data, without checking the resulting header: make sure that you understand what each keyword in means, and that it is still applicable (e.g. 2dfdr included 2D WCS metadata, providing information about the wavelength-pixel mapping, and the fibre-pixel mapping, if you copy across the header when you extract a single spectra, the second WCS axis is logically invalid).

  3. Specify all the units within the metadata, via standard keywords such as BUNIT and CUNITn. This means Data Central does not need to track units separately, and that standard tools (such as astropy) can display, convert and manipulate data without users needing to manage units manually. See Units within Data Central for some unit-specific advice, which applies across all of the different data types and formats (including catalogues).

  4. Specify bibliographic and provenance information in the primary header, via standard keywords such ORIGIN and REFERENC (please include as much as possible, even down to the specific data release the file is from). Whilst Data Central does not specifically use this information, tools such as specutils can use this to identify and load your files, and the files can understood outside the context of Data Central.

  5. Don’t include images/spectra/other data from other surveys within the same FITS file—this results in duplication of effort, increases the amount of code needed to manage these files (e.g. custom readers have to be written), and wastes storage and bandwidth. Instead, include references to other surveys via the survey object name (this is best done via catalogues, however this can additionally be stored in metadata within headers of the relevant files). The VO (Virtual Observatory) tools are designed for this kind of work, and future iterations of Data Central will expose more and more of these inter-survey links, and hence increase the richness of the visualisations.

  6. Include metadata about the different HDUs within the FITS file. Use standard keywords (like EXTNAME) where possible, but there are other non-standard keywords such as ROWn or ARRAYn which have become unofficial conventions.

A reference containing all the standard FITS keywords can be found at https://heasarc.gsfc.nasa.gov/docs/fcg/standard_dict.html and further links to other sets of keywords and conventions can be found at https://fits.gsfc.nasa.gov/fits_standard.html.

Imaging

Coming soon.

IFS

cube_blue/cube_red

Coming soon.

Spectra

spectrum_1d

1D Spectral FITS files must have a single source spectrum in a single row. Flux, SN, sky emission can be included in different FITS extensions.

If you are registering a 1D spectra in your <survey>_<datarelease>_product_meta.txt, you must also include the following columns in the <survey>_<datarelease>_product_meta.txt file:

Column

Description

fluxHDU

Number of extensions containing source flux

snHDU

Number of extensions containing source signal to noise (can be blank if not provided)

skyHDU

Number of extensions containing sky emission (can be blank if not provided)

fluxUnitKeyword

FITS header keyword containing flux units (can be “ergs/sec/cm^2/ang”, “ergs/sec/cm^2/Hz”, ‘jansky”, ‘counts’)

fluxScaleKeyword

FITS header keyword containing scaling for flux units (numeric)

wavelengthUnitKeyword

FITS header keyword containing wavelength units (can be ‘ang’, ’nm’, ‘micron’, ‘m’)

CRPIXKeyword

FITS header keyword containing pixel reference point for wavelength scale.

CRVALKeyword

FITS header keyword containing pixel reference point value for wavelength scale.

CDELTKeyword

FITS header keyword containing pixel width for wavelength scale.

spectrum_2d

Coming soon.

Metadata File Formats

Metadata files are always pipe-delimited, and have the extension .txt

Documentation

If you would like formatted text to appear in the Schema Browser, you can supply a “documentation” entry for groups and tables. The file itself can contain html text.

Please use the following extensions (to denote the format of a particular documentation file):

.html contains html-formatted documentation

.txt contains plain text-formatted documentation

If you do not wish to supply documentation for a particular row, leave that entry blank.

If you would like to include email addresses in your documentation, please format as:

<a class="is-mailable" data-person="john.smith" data-domain="sydney.edu.au"></a>

This is an anti-spam measure: rather than placing raw html (easily harvested by spambots), a javascript-enabled browser is required to decode the address.

Documentation that you will want to update frequently should be maintained by the survey teams themselves in Data Central’s Documentation portal: Document Central. Contact us to get your team set up within the system if you have not done so already.

Important

Remember that the documentation used in the ingestion is the static, paper-like documentation, the documentation on Documentent Central is entirely separate.

Units within Data Central

Units and their representation is one of the more complex parts of the metadata Data Central tracks, due to the existence of differing, contradictory standards for representing units, and a large legacy of existing data which does not match a particular standard chosen. Data Central uses standard tools (such as astropy) to read, convert, write and generally handle units in a VO-compliant way as possible, and aim to fix all unit issues before any public data release. In general, Data Central will not modify existing data releases to fix issues with units.

In terms of actually representing units, we suggest reading VOUnits, the IVOA Recommendation on units within the Virtual Observatory, as it provides useful information about how existing unit standards (or conventions), both within astronomy and outside, specify how to encode units. The VOUnits document has nothing particularly unusual, though it does note some corner cases worth taking note of, such as treating the string 'Pa' as Pascals rather than petayears (which would happen if naively if 'P' was always treated as representing Peta).

In general, there is nothing particularly strange you need to do to produce data with VO-compliant units, but there are two things you should probably check: the capitalisation and pluralisation of the units; and encoding dimensionless or unitless quantities.

Capitalisation and Pluralisation of Units

Astropy by default reads only the singular form of a unit (e.g. “second” over “seconds” or “Angstrom” over “Angstroms”), as generally this is the preferred form in most unit standards. Data Central handles plural forms for existing data releases, but for maximum compatibility, the singular form should be used for new data (we are unaware of any software which rejects the singular form).

Similarly, lower case (except for units named after a person) is to be used, with capitalisation of the first letter if the unit is named after a person (e.g. “Angstrom” over “angstrom”). All caps (every letter being upper case) will throw errors in astropy, and therefore should be avoided in new data.

Encoding Dimensionless or Unitless Quantities

Neither the FITS standard nor the IAU Style Guide discuss the representation of dimensionless/unitless quantities, so there is no specific representation in standard FITS. There are some differing conventions mentioned in the VOUnits document, with the VOUnits document recommending using an empty string.

For normalised quantities, such as normalised intensity, we suggest using the empty string for the unit, and describe the quantity in the comment (e.g. have the FITS comment for the field be “Normalised Intensity” for something stored as normalised intensity).

Processes

If you have any questions about this documentation during the preparation of your data release, please contact us.

Response Timescales

We endeavour to implement any change-requests as soon as possible, however the DC team is currently relatively small. Please see below for a rough estimate on update timescales. Please note, these are a guide only, subject to the complexity of the update requested, and additional constraints on our time (e.g., other survey-team data releases, conferences, funding milestones etc).

Data and Schema Changes

For critical changes (e.g., data product files/catalogues need to be replaced with new versions), we will work towards an update turnaround of ~1 week. For non-critical changes (e.g., minor schema changes) we guarantee update timescales less than 1 month. To clarify whether a change is critical or non-critical, contact us.

Interface Issue

For critical interface issues (e.g., blocking data access) we aim to address any issue submitted by any user within 1 week (though usually within 48 hours). For non-critical issues (e.g., mis-alignment on mobile view) we aim to publish a fix within 1 month.

Feature Requests

Data Central is developed with its users in mind. We’d love to hear your suggestions for new features you’d like to see developed, plus give you the opportunity to vote on the functionality you’d like to see prioritized.

Please suggest new features and vote on existing suggestions on our requests page. We use this information to guide our milestones and future development effort.

Policies

We reserve the right at any time to remove data that is in violation of any data usage licenses or agreements.

Data

When you release data through Data Central, all data and meta data are available for public download and distribution. Unless you request otherwise, data will be publicly released with a Creative Commons Attribution 4.0 International (also known as CC BY 4.0). See https://creativecommons.org/licenses/by/4.0/ for more details.

Who owns the data?

You do! Any data that is uploaded to Data Central belongs to the researchers who created it. We are merely custodians of your data, allowing you to share it publicly or with your team.

What data can we share?

We will only share public data and only with the owner’s permission.

With whom can we share data?

If data has been made public, then anyone has access to it. Private team data is only available to team members, and access is in the control of the team.

Maintenance

We aim to push new code to the production server on a regular basis and as such, guarantee a quick turnaround for bug fixes and data updates. Downtime and maintenance windows are currently unscheduled, but we aim to keep disruption at a minimum (i.e., <1h). You can check our twitter for maintenance notices @DataCentralOrg.

License

Copyright 2019 Data Central

Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the “Software”), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:

The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.

THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.

Support

Got a question? You can email us at contact@datacentral.org.au or send us a message via our support site.