Difference between revisions of "Help:Import Spreadsheet"

From DAS Wiki
Line 15: Line 15:
  
 
==The Table Formats==
 
==The Table Formats==
Note that you do not have to have an input for every field. You may not have any entries in any of the records for a particular field. However, the fields that you do have must be named (in the top row of the CSV file) ''exactly'' the field names below, including underscores, if any. And the names are case sensitive.
+
Note that you do not have to have an input for every field. You might not have any entries in ''any'' of the records for a particular field. However, the fields that you do have must be named (in the top row of the CSV file) ''exactly'' the field names below, including underscores, if any. And the names are case sensitive.
 
===The Arch Table===
 
===The Arch Table===
 
{|class="wikitable"
 
{|class="wikitable"

Revision as of 17:02, June 29, 2021

We know that many researchers have used spreadsheets to organize their research, and it would be a waste of time if, in bringing that data into the wiki, it had to be entered all in by hand again. Fortunately the wiki has an import function where properly-formatted spreadsheet data can be automatically imported into the wiki.

CSV Files

In general, the process requires that the spreadsheet be exported to something called a comma-separated-values file (or CSV, for short). Each row in that file will correspond to a future page on the wiki, and will be populated with whatever fields of data you have that correspond to the wiki fields. Nearly all spreadsheet programs have an "export to CSV" option, so the export to that format should be easy.

However, the file should be prepared in a particular format prior to export. Here are general considerations:

  • Any fields that contain geographic coordinates (lat/long) must be decimal-degree values, and in a string " " separated by a comma, i.e "39.2356,-74.2356" otherwise they'll import, but won't be the right location. Note that decimal degrees is not the same as degrees, minutes and seconds. See the difference: 30.5 degrees (which is decimal notation) is the equivalent of 30°30’00” (degrees, minutes, seconds notation). Do not take something looking like 30°23’46” and enter it in as 30.2346. They are not equivalent and you'll find the location will not be where you intend.
  • Any commas in any of the fields must be within quotes. For instance, brick, masonry, stone MUST be "brick, masonry, stone" or the import will fail.
  • Make sure there isn't already a page for the record in the wiki. If there is, delete that record from the CSV and enter its data into the wiki by hand.
  • The header record (field names, etc) which is the first line of the CSV file has a particular format requirement. The page field MUST be named Title and the rest of the fields MUST be named Table_Name[field_name] (Table_name to be substituted for whatever table the spreadsheet is to be imported into. Currently we have Arch, buildings, Firms and Grounds.)

The Table Formats

Note that you do not have to have an input for every field. You might not have any entries in any of the records for a particular field. However, the fields that you do have must be named (in the top row of the CSV file) exactly the field names below, including underscores, if any. And the names are case sensitive.

The Arch Table

Field Name Type Comments
portrait File
birth_date Date
birth_approx Boolean
birth_place String
death_date Date
death_approx Boolean
death_place String
nationality String
alma_mater String
associated_firms List of Page
burial_place String
spouse String
signature File
awards String
significant_design String

The buildings Table

Field Name Type Comments
image File
address String
Geo Coordinates
altitude String
owner String
building_type String
architectural_style String
height Float
material String
floor_count Float
floor_area Float
elevator_count Integer
structural_system String
architect Page
architecture_firm Page
structural_engineer String
civil_engineer String
m_e_engineer String
other_designers String
surveyor String
landscape_architect String
contractors List of Page
awards String
groundbreaking_date Date
groundbreaking_date_approx Boolean
start_date Date
start_date_approx Boolean
topped_out_date Date
topped_out_date_approx Boolean
completion_date Date
completion_date_approx Boolean
opened_date Date
opened_date_approx Boolean
cost Float
cost_approx Boolean
renovation_date Date
renovation_date_approx Boolean
ren_architect String
ren_firm String
ren_l_a String
ren_str_engineer String
ren_civ_engineer String
ren_oth_designers String
ren_m_e_engineer String
ren_contractors List of Page
ren_cost Float
ren_cost_approx Boolean
ren_awards String
demolished Date
demolished_approx Boolean
national_register_id Integer
national_register_url URL

The Firms Table

Field Name Type Comments
logo File
image File
firm_type List of String
founders List of Page
antecedents List of Page
successors List of Page
personnel List of Page
location String
founded Start date
founded_approx Boolean
dissolved End date
dissolved_approx Boolean
awards List of String
significant_projects List of Page
website URL

The Grounds Table

Field Name Type Comments
alt_name String
image File
address String
perimeter Text
Geo Coordinates
founder List of String
owner String
builder List of String
landscape_architect Page
other_designers List of Page
acreage Float
elevation Float
founded Date
founded_approx Boolean
opened Date
opened_approx Boolean
type List of String
NRHP_ID String
NRHP_url URL

The Import

After the CSV file is in a format that allows for one-to-one import to the wiki fields, click on the Import CSV page, select the CSV file using the "Browse" button at the top of the page, leave the encoding type as-is, select "Skip" for pages that already exist, and click on the Import button at the bottom to initiate the process.

Note that wikis do not import files of this type all at one time. To lessen the load on the site, they import one record each time there is another page request anywhere else on the wiki. This serves to spread the backend processing load out and allows the wiki to remain responsive to other users. Over the next few minutes or hours the wiki will import one record at a time from the CSV until all the records are online.