Home
User Manual aimed at end-users
Contents
1. A e j i umee pipat hean ip ee eg ee eee gg ee eee i se MS ae ee oe T kemi Mie Step 3 5 5 1 1 Inventory Step 4 5 5 1 2 Process inventoried Step 5 5 5 1 3 Harvest Step 6 5 5 1 4 Process harvested Step 7 5 6 1 1 Synchronise Step 8 5 6 1 2 Extract e NNN Nm The order of these operations is important and for all protocols DiGIR BioCASe and TAPIR each operation must be run sequentially in the order above 5 3 Registry Synchronisation Synchronising with the GBIF Registry populates a list of Metadata Updater BioDatasources from the list of available endorsed access points that are registered In the Registry you have Organisations that represent data publishers Resources that represent datasources and Services that represent access points The HIT creates a new Metadata Updater BioDatasource for each service and saves it to the bio_datasource table 5 3 1 How to run the operation s 5 3 1 1 Synchronise with Registry e Navigate to the Registry tab e Choose which Node you would like to synchronise for default is all Nodes and then click Synchronise See picture below 5 3 2 How to monitor the operation e Navigate to the Console tab e Until the synchronisation finishes log messages will display information about newly created updated BioDatasources files containing provider contact information that are written as well as any error or warning messages that should be brought
2. and Process Harvested Each of these operations is explained separately executing them one after the other in the exact order they appear e USER TAKE NOTE Usually for new data publishers it s a good idea to run through the operations separately carefully watching the responses and generated files for any irregularities For well behaving data publishers it s much more convenient to dispatch the operations in bulk for example dispatching an inventory amp processInventoried amp harvest amp processHarvested at the same time 5 5 1 1 Inventory ed A A R Tera mem p re ee ber pa o m mmeg ai meg eee e m ERECT L E ood ir on ee ee eee oy Leri NA Driana Cmr ii mai bani mi m on 9 he ieee 1am sm Te ee BOLETI Bin oe a re miep oi hei e i a a Gima i ee es aii T ee er a Bj i ae Ta d Ribs Si idiki n N E E a a L a i rrn Le m a ee re be r a an ee SSS eme aa E In this operation an Operator BioDatasource requests a list of all the scientific names that are contained in the dataset resource that it corresponds to In the same request it also requests information about the count of each scientific name that is the number of unique records that exist for each scientific name The outputs from this operation are 1 One or more inventory requests see picture below AHL a ere le oe ee ho EE oe ec ae et aheekr iii 1 5 D see wien Slate Pon cers The AALS rip Hrd Te A Vihar idem friaa Hirik y e
3. scientific names in the inventoried file will contain no duplicates and arrange the scientific names alphabetically If the list does not have these characteristics double check the inventory response s to ensure that the names are in fact returned in order 5 5 1 2 Process inventoried ss mm e D goa m poa kase j beite al iei bee tamsi mm e In this operation an Operator BioDatasource reads the file outputted from the inventory operation inventoried txt and generates name ranges composed of lower and upper scientific names The range is later used to restrict the search requests sent during the Harvest operation This strategy of requesting records by name ranges has been found to be the most efficient way of harvesting all of a data publisher s records This is explained in more detail in the Harvest section following below The outputs from this operation is A single file containing the list of scientific names and their counts inventoried txt see picture below Crececcre T FEF ee ALENT EL TT HER Lan ds salle Hi EEERERTERECTOT CELT Tee FLGERLLPECESCEESS 2 A single file containing all the name ranges that were contructed nameRanges txt Wherever possible the size of the name range that is the number of records that it represents is included beside each name range see picture below K an E P A ERG i tee awl I ee ee ra eee tt USER TAKE NOTE With TAPIR and DiGIR it s possib
4. the Operations and Procedures Metadata Updater section 5 4 4 Some additional information e Imagine that any metadata that you d like to synchronise with your indexing database would get collected at this stage Currently the application is configured to harvest only that metadata that fits into the GBIF portal database s data structure For information on how to modify the metadta that gets harvested the process is quite simple and information on how to do so can be found on the Extensions and Customisations Modifying a Metadata Mapping section 5 5 Harvesting Harvesting is the act of act of gathering records from data publishers Harvested information is gathered and stored as raw XML and then processed into intermediary text files which are later used in indexing Harvesting always operates on individual data resources with statistics and log messages presented to the user so that they can better manage their network of data publishers 5 5 1 How to run the operation s For brevity all the operations mentioned in this section are run in a similar fashion to the Metadata Update operation except that they are run via a Operator BioDatasource tick the checkbox amp schedule In one go the user can schedule any number of operations across any number of BioDatasources at the same time There are four operations relating to harvesting UserManual5 5 1 1_ Inventory Inventory Process Inventoried UserManual5 5 1 3 Harvest Harvest
5. KE NOTE 1 In terms of the information that gets stored in the GBIF Portal database a core record can also have multiple images identifiers typifications or links associated to it Here is a breakdown of the various tables that could get populated during synchronisation depending on the information that s associated to each record portal raw_occurrence_record core record information i e scientific name country etc portal image_record multimedia information associated to a core raw_occurrence_record record i e an image URL portal identifier_record identifier information associated to a core raw_occurrence_record record i e a GUID or LSID portal typification_record typification information associated to a core raw_occurrence_record record i e its type status portal link_record links associated to a core raw_occurrence_record record i e some species page URL 5 6 1 2 Extract In this operation an Operator BioDatasource will extract the raw records that were stored during the synchronisation operation performing further data quality validation and classifying the record against the GBIF nub taxonomy please see this page for more information on the GBIF nub taxonomy Successfully passing through this stage a record ends up in a state where it s ready for accession via the GBIF Portal The data quality checks taking place here include longitude latitude coordinates are checked against the specified country to e
6. X __ gbif indexingtoolkit ru meow emanas aur Ne GBIF Harvesting and Indexing Toolkit HIT Project Home Downloads Wiki Issues Source Export to GitHub Search Current pages for Search UserManual User Manual aimed at end users Featured Phase Support 1 0 Background Information Available here 2 0 General Specifications Available here 3 0 Installation and Configuration Available here 4 0 Introduction for First Time Users My favorites w Sign in Search projects Updated Nov 8 2010 by kyle br qmail com Before any harvesting and indexing H amp l can take place please ensure that you have properly installed and configured your installation Although it is not completely necessary some knowledge of the various harvesting protocols would be valuable For more information on each of the supported protocols you can check the Supporting Information on Protocols page In this manual each of the HIT s different operations is described in brief For a more technical description you could refer to this page Ultimately it is only when things go wrong during debugging that an intimate knowlege of the protocols would be beneficial Otherwise expert knowledge of how the HIT works i e what requests are being sent what responses are expected the files that should be written and so on will prove most useful as the HIT hides the nitty gritty details from the user presenting the same set of opera
7. ess_point table It is in this way that each contact is assigned a unique agent identifier each data publisher is assigned a unique data_provider identifier each data resource is assigned a unique data_resource identifier and each data_resource s access point url is assigned a unique unique resource_access_point identifier gt Sign in to add a comment Terms Privacy Project Hosting Help Powered by Google Project Hosting
8. i e the machine on which the HIT is running goes down Picture below shows output log messages from a successful harvest operation Lege seer Lies e USER TAKE NOTE 1 For large datasets the harvest can take an extremely long time For example a TAPIR dataset having 3 million records would take a week to complete varying of course according to connectivity speed etc e USER TAKE NOTE 2 The user can monitor all harvest operations which have been started and are still running by paying attention to the Harvested Start column in the BioDatasources list on the Datasources page See picture below e DEBUGGING HINT 1 The most common reason that a search response is invalid is that it contains an XML breaking character When a name range representing 500 records fails for example it could be due to a single invalid record and as a result the other 499 records do not get harvested In an effort to harvest as many records as possible and help the user identify where the breaking characters are found the system will break a request that fails into several smaller requests Keep a careful eye on the output log messages for which responses are invalid and provide feedback to the data publisher which will help them improve the quality of their dataset e DEBUGGING HINT 2 With thousands of output log messages being generated sometimes it s more convenient to look at a BioDatasource s rolling log file for aggregated view of its activi
9. iir aaia epii TR fees Ae ieee 2 One or more inventory responses see picture below P oat ele mrri ria ae Omit hem r EA l O T ROO Cael T r umer mem ee Harari le oga rip Lpi ai brp Erbe La E eee a A e Be a Ey AT i Oren eei Rr ees rd ee pe ed Lees eb a eee oiler a ha birik Vam rem rere tarry bes Cees fear ba Serer heer diar hoe ama re ee Py ee eee bes Se be tr he eee eee pe pardalar ee be eee Picture below shows output log messages from a successful inventory operation Leg E erel Lal me su by Lemme Bmg pm ma el pm wee me ban el ba ol i jh mem al Se a cee eal Pir k Be Fe Ba ari i mer eres it ee ee i ae a Jeers p see Din Ba Of ad bekri fuar ere hoe il reie eB eed ge m PPa b m PS mi p oei miem gi ae mii ri Sedai Fe ps Legere j kmp F erea e ME bT npe Shei doi a i ie LT he yy ey ee 7 a week i ree ee ee ae ee ee ee Sara ey i iii fare i ie LEAI F ETE ee a Se SAA s ihj feel DE ie Eam E N A Fip b M pa i ek ia T bime i iy ap P b aea ee E R i n e USER TAKE NOTE For BioCASe no count information is ever collected e DEBUGGING HINT 1 Often the reason no inventoried list could be generated is because the inventory response was empty From the Console tab the xml requests and responses can be checked directly from the browser e DEBUGGING HINT 2 The integrity of the inventoried list is paramount to the success of subsequent harvesting operations Ideally the list of
10. int such as can be the case with DiGIR and BioCASe they all get converted into separate Operator BioDatasource which can be harvested independently Information such as descriptions record counts and contact information is gathered and saved for each Operator BioDatasource The whole of this work is carried out by a single operation Metadata Update 5 4 1 How to run the operation s 5 4 1 1 Metadata Update Navigate to the Datasources tab and filter for a particular Metadata Updater BioDatasource that you are interested in harvesting See the picture below using the provider Academy of Natural Sciences as an example F EF TET HNT TG EI Pe OT Ee Eee ee Beau Lint OEE es eri p er eee ed ed ee ee eee ee pe See eed ee ed prire ee hr ee am ee es eee eg ee b bees s Mma bece onm mm ye ae ig ae me pie Se e m X lt p D a r gt D W O O O n O C Q D 09 D D O O a Cc D o D O xr r 2 iai or since Fes e Click on the operation Metadata Update and then click the button Schedule to submit the job See picture below Ea ar CD HARSA A a ce Hoiu Lii L E EESAN EN eee TA EEN PE P ee ee EA cee mel be oo itis oP ea Poe ty Tet LO Cie ee oe an ee a SS a a ee A a As a ee fe u n 5 4 2 How to monitor the operation s e Navigate to the Jobs tab and see that a job for the metadata update has in fact been created If you re too sl
11. le for publishers to specify an attribute in their metadata that limits the number of records that can be requested for at one time The value of this attribute is taken into account when determining the size of name ranges Otherwise the system default name ranges to 200 records for both TAPIR and BioCASe and 900 records for DiGIR DEBUGGING HINT 1 The only reason that the name ranges file couldn t be generated is if the inventoried list of scientific names was empty or all scientific names were invalid Note that scientific names containing SQL breaking characters such as amp are still included but the breaking characters are replaced automatically Therefore at this level there is a data quality check on scientific names and any errors are outputted as log messages to the Console DEBUGGING HINT 2 Often the reason a harvest does not retrieve 100 of a dataset resource s records is that not all records are covered by the name ranges that have been generated From the expanded BioDatasource in the BioDatasources list in the Datasources tab you can view the name ranges file directly from within the browser Compare this file against the inventoried txt file for any inconsistencies 5 5 1 3 Harvest ai O e eee a ATEA In this operation an Operator BioDatasource reads the name ranges file outputted from the Process Inventoried operation nameRanges txt and dispatches a search request for all records greater than or equal to the lowe
12. ment 2 identifier_records txt a tab delimited text file containing a header line with column names with each line representing an identifier record i e GUID relating to a given Unit record 3 identification _records txt a tab delimited text file containing a header line with column names with each line representing an Identification element relating to a given Unit record element 4 higher_taxon_records txt a tab delimited text file containing a header line with column names with each line representing higher taxon elements relating to an Identification element which ultimately relates to some Unit record element 5 link_records txt a tab delimited text file containing a header line with column names with each line representing a link record i e URL relating to a given Unit record element 6 typification_records txt a tab delimited text file containing a header line with column names with each line representing a typification record i e type status relating to a given Unit record element USER TAKE NOTE 1 Out of the box the HIT is configured for parsing only those elements that are of interest to GBIF or in other words that fit into the GBIF Portal data structure If wanting to index into a different data structure and thus parse a different set of elements of interest you could refer to the sections Adding a New Synchroniser and Modifying an Index Mapping USER TAKE NOTE 2 The user can monitor all harve
13. nsure that the record is georeferenced accurately the scientific name and taxonomic information are all validated in the aforementioned classification e the basis of record name is validated against a controlled vocabulary e etc Picture below shows output log messages from a successful extract operation USER TAKE NOTE Here is a breakdown of the various tables that could get populated during extraction depending of course on the information that has been stored for each raw record e portal occurrence_record core record information having been classified now against the GBIF nub taxonomy 5 6 2 How to monitor the operation s While synchronisation runs log messages are being output to the HIT s Console tab so that the process can be monitored Note that for very large data resources synchronisation can take quite a while approximately 1 million records hour In addition log messages are also being written to a table in the GBIF Portal database called portal gbif_log_ message Such log messages are those that will appear in the GBIF Portal for example and provide back information to the data publisher about the success of the attempt to index their data resource 5 6 3 How to debug the operation s if necessary If information is not being synchronised with the database correctly it is suggested that you check e Did the data get harvested correctly into the intermediate text file s generated during the Process Har
14. ow though the job might have run and finished in which case it will disappear from the Jobs list See picture below F ie ae eT es os PT woh Loa Se el ees ees eed ee es eee Se er 2 ee ey ere oe ee a ire co m 7 ee es ere ey a nr nae e From the expanded BioDatasource click on the link Logs which will take you to the Console tab with output log messages for only this BioDatasource This is preferable to referring to the Console tab when several different jobs are running concurrently and all output log messages will get written to this page in the order they are written See picture below e The end result of a successful metadata update is the creation or update in the case that they already existed of one or more Operator BioDatasource Filter the BioDatasource list once again by the name of the provider for example in the Datasources tab to see the list of Operator BioDatasource that are now available to harvest See picture below 5 4 3 How to debug the operation s if necessary e If the job metadata update has completed there are several problems that might have occurred In most cases the output log messages will be descriptive enough to inform you what interrupted a successful metadata update Note that behind the scenes the activity can vary greatly between protocols For a descriptive account of how metadata updates are carried out for each of the different protocols you could refer to
15. r name range and less than or equal to the upper name range Where a single request fails to bring back all records for a particular name range the system automatically generates additional requests asking for the next records until all records for that range have been harvested Name ranges are iterated over one after the other from start to finish until all name ranges have been covered It is in this way that a data publisher s whole dataset resource gets harvested The outputs from this operation are 1 One or more search requests with enumerated extensions corresponding to the order in which they were dispatched i e search_request 000 2 One or more search responses with enumerated extensions corresponding to the order in which they were dispatched i e search_response 000 Often there will only be a single response per request but sometimes there can be multiple responses for a single request While this operation runs two files are constantly being generated and regenerated 1 failedNameRanges txt maintains a list of all name ranges that couldn t finish successfully The existence of this file allows the system to re run the harvest particularly for those ranges that have failed and need extra attention 2 pendingNameRanges txt maintains a list of all name ranges that have yet to be read The existence of this file allows the system to pick up where it left off in the event that harvesting gets interrupted for some reason
16. rocess harvested o 5 5 2 How to monitor the operation o 5 5 3 How to debug the operation s o 5 5 4 How to monitor the operation o 5 5 4 Some additional information e 5 6 Indexing o 5 6 1 How to run the operation s m 5 6 1 1 Synchronise m 5 6 1 2 Extract o 5 6 2 How to monitor the operation o 5 6 3 How to debug the operation s o 5 6 4 How to monitor the operation o 5 6 4 Some additional information 5 2 Quick Start Guide Once a Step 1 5 3 1 1 Synchronise with Registry has been done with the GBIF Registry Mm SaaS as Ha DEDS HAL AEI AAU DOLORI PRT rT Regain a a ak I a I Ee a aie RS A a aiai i E pE ee eee bo eg ee and the full list of endorsed and available access points have been registered as Metadata Updater BioDatasources the user must perform a Step 2 5 4 1 1 Metadata Update against one or more access points of interest P m D is i kaion mi Be pe This discovers all the available datasets that are available behind a given access point and registers them allas Operator BioDatasource Then for one or more datasets of interest you can begin to harvest and index them running through the following operations in order K me C HAEATA AH PRAD LTI Bhandari ee mee fe be ome baam eg i a cage meme ea os a cpa Kalama eins pel ois eine winced Do im a e pars prt rA A p E ET U E a EES E AE E A e pert e e rra e i a e pe ee ai BHOCGLE RS SGsR RRP ETS iE J DEH es LE SD eS OS YA A bme ij Gare I
17. st operations which have completed by paying attention to the Last Harvested column in the BioDatasources list on the Datasources page In addition the Dropped Harvested and Max Harvested columns will get updated reflecting the harvesting statistics See picture below DETTE i i q i DEBUGGING HINT 1 Simply put the reason no or not all records are processed and the output harvested file s are empty or incomplete is that not all response files could be processed or they didn t actually contain any records The reasons for this may vary greatly and you should analyse the output log messages on the Console tab carefully 5 5 2 How to monitor the operation s Similar to advice that has been reiterated before check the Console tab filtered by BioDatasource 5 5 3 How to debug the operation s if necessary The order of operations is important And because one operation usually depends on one preceding it it s usually best practice to start checking for problems from the bottome up first with the inventory then the process inventory then the harvest and then the process harvested 5 5 4 Some additional information 5 6 Indexing Indexing is the act of synchronising the processed harvested records stored as intermediary text files with a database and performing the necessary data quality checks and processing to get the information into an appropriate format Out of the box the HIT is capable of indexing in
18. tions to the user regardless of which protocol is being used Once you re ready to begin it s recommended that you step through each of the following sections slowly while carefully observing that things are running smoothly For testing it s important to pick a data publisher that you re confident works and that publishes using a protocol and encoding that the HIT is capable of supporting See the Specifications page for information on those protocols and encodings that are supported 5 0 User Instructions 5 1 Overview In essence there are four major sections that group one or several separate operations together Each section is then divided into four additional parts How to run the operation s How to monitor the operation s How to debug the operation s if necessary Some additional information oe o e 5 3 Registry Synchronisation o 5 3 1 How to run the operation s a 5 3 1 1 Synchronise with Regist o 5 3 2 How to monitor the operation o 5 3 3 How to debug the operation s o 5 3 4 How to monitor the operation o 5 3 5 Some additional information e 5 4 Metadata Update o 5 4 1 How to run the operation s a 5 4 1 1 Metadata Update o 5 4 2 How to monitor the operation o 5 4 3 How to debug the operation s o 5 4 4 How to monitor the operation o 5 4 4 Some additional information e 5 5 Harvesting o 5 5 1 How to run the operation s a 5 5 1 1 Invento a 5 5 1 2 Process inventoried 5 5 1 3 Harvest m 5 5 1 4 P
19. to the GBIF Portal data structure The whole indexing process is divided into two parts which will be explained below Synchronise and Extract 5 6 1 How to run the operation s For brevity all the operations mentioned in this section are run in a similar fashion to the Metadata Update operation except that they are run via a Operator BioDatasource tick the checkbox amp schedule 5 6 1 1 Synchronise keye ee ce ee ey i MART a a M mem r P b be me a In this operation an Operator BioDatasource reads the harvested record file s output from the Process Harvested operation For each core record the system checks whether it already exists or not The check for uniqueness is based on three required fields that each record must have collection code institution code and http rs tdwg org dwc terms catalogNumber catalog number If the record does not exist it gets added If the record is found to exist however the corresponding record is updated with the new record s information At this level there is some data quality checking taking place Namely e date fields such as the collection date and identification date will get will get validated existence of none multiple preferred or multiple unpreferred higher taxonomic identifications will get flagged e usage of an invalid type status will be ignored e etc Picture below shows output log messages from a successful synchronise operation e USER TA
20. to the user s attention See picture below 5 3 3 How to debug the operation s e When synchronisation with the Registry finishes all the Metadata Updater BioDatasource that have been created should be available from the Dashboard Datasources tab This type of BioDatasource is colored orange to differentiate them from the Operator BioDatasource which is colored green See picture below Note that no Operator BioDatasource will appear at this stage as they only get created following a metadata update which is explained in the next section e If there is a particular access point that hasn t been converted into a Metadata Updater BioDatasource then you could check the output log messages in the Console tab for some explanation e USER TAKE NOTE Care should be taken when deleting BioDatasources If a Metadata Updater BioDatasource gets deleted it is flagged in the HIT database as deleted and will not be re synchronised again To see a list of all BioDatasources that have been deleted the user can click on the deleted link on the Dashboard See picture below TERE SEREE J r bhed e mnai i mani E waas e im 5 3 4 Some additional information e Please visit this page http gbrds gbif org to explore the GBIF Registry 5 4 Metadata Update A Metadata Updater BioDatasource asks for all the metadata behind the given access point that it corresponds to If there are multiple data resources behind a given access po
21. ty This file is available from a link underneath the available options in an expanded BioDatasource in the BioDatasources list 5 5 1 4 Process harvested records In this operation an Operator BioDatasource will collect all the search responses parse them and write the parsed values to file Mapping files identify which XML elements are to be parsed The file s written are then used during indexing to synchronise the harvested records with the database The outputs from this operation depend on which encoding was used to wrap the response If the response was encoded using DwC there is a single output file 1 harvested txt a flat tab delimited text file containing a header line with column names See picture below nian ee CUCU p TEEI aT mI ME TIN HEHNSHN J 1 TEETER TRB ENEE N TOOT TET eT reeled rele TATE TEE HE TATE HHRHHS If the response was encoded using ABCD there is one core file 1 unit_records txt a tab delimited text file containing a header line with column names with each line representing a single Unit record element and six additional files all relating back to the core file and thus handling for the hierarchical structure of ABCD responses versus the flat structure of DwC responses 1 image_records txt a tab delimited text file containing a header line with column names with each line representing a multimedia record relating to a given Unit record ele
22. vested operation e f some data is misrepresented i e it falls into the wrong columns in the table you could check to see that the right index mapping file is getting used and that its mappings are in fact correct e f some record s term concept isn t getting synchronised at all with the raw_occurrence_record table ensure again that the correct index mapping file is being used and that the term concept of interest is mapped and that the corresponding XPath is correct used to parse the XML response for that term concept If information is getting synchronised with the database correctly but it is not getting exposed in the Portal sitting atop the GBIF Portal database it is suggested you check the indexing history page available by data resource in the Portal 5 6 4 Some additional information Harvested records information is not the only information that gets indexed At the beginning of indexing the following metadata also gets synchronised e data publisher contact information synchronised with the portal agent and portal data_provider_agent tables e data resource contact information synchronised with the portal agent and portal data_resource_agent tables e data publisher information synchronised with the portal data_provider table e data resource information sycnchronised with the portal data_resource table e additional resource access point information for that data resource synchronised with the portal resource_acc
Download Pdf Manuals
Related Search
Related Contents
TECHCONNECT HDBT MANUAL DEL PROPIETARIO Copyright © All rights reserved.
Failed to retrieve file