Aleph MAB
This mapping to the normalized record is based on the internal structure of MAB2 within Aleph 500. The description of this mapping uses the subfield structure of Aleph.
Indicator/Subfield:
-
The '#' symbol indicates the truncation sign for one or more indicators(s)/subfield(s).
-
The '$$' symbol indicates the subfield (followed by the subfield code).
-
The '_ 'symbol indicates a blank indicator.
If multiple fields and/or different fields are mentioned in the Source column of each mapping section, all fields are taken as sources for the PNX field (= and). Otherwise, the "OR" term indicates that the following field should be taken if the previous field cannot be mapped or does not exist).
Rules for Parent-Child Relationship
The MAB2 relationship between multipart works (record type "h") and volume records (record type "u") is referred to as a parent-child relationship.
There is a general rule concerning the integration of "parent" information into a "child" record during the extract process: All data from the parent expands to the child, including subjects and classifications.
The expand program (expand_doc_mab_recursive: child - parent, if 010 is present) is used in the extract process. This program, by default, always adds the second indicator to all fields of the record itself. So the second indicator with value of 1 is present in all records, not only for child records, and the fields from the parent are added using the current convention of using the second indicator for the level: 1 = child/present record and 2 = parent). The second indicator defines which fields to use from the present records and from the parent. This is added to the mapping below. There are four general rules concerning the mapping from MAB2 to the fields of the PNX:
Rule 1
Take only field(s) of the parent record OR only field(s) of the present record (in these cases the second indicator 1 or 2 is mentioned explicitly)
-
field information from just one record
Rule 2
Take field(s) of the parent record ADD field(s) of the present record (in these cases the second indicator is mentioned as #)
-
full integration of parent field information
All Aleph MAB fields of the parent are taken into the PNX field (even if this field exists in the child (for example, for creators, titles, subjects) and then all Aleph MAB fields of the child (the rule for the first indicator is mentioned in the Source column). The merging rules for the sequence/merge of the taken fields are mentioned in the Note column (mainly according to fields of the same record, sometimes grouping according to field names).
Rule 3
Take first fields of the present record OR fields of the parent record if the field is missing in the present record; take into account the EXACT FIRST INDICATOR for the integration of field information
-
partial integration of parent information WITH consideration of EXACT FIRST INDICATOR
Rule 4
Take first field(s) of the present record OR fields(s) of the parent record if the field is missing in the present record; do not take into account the first indicator for the integration of field information.
-
partial integration of parent information WITHOUT consideration of exact first INDICATORFor Rules 3 and 4, if various source fields are mentioned in the table (Source column), apply the rules field for field. For example, if you apply Rule 4 to the following: Field AAA##, Field BBB##.
-
Take child field AAA##. If not existing, take the parent field AAA##.
-
Take child field BBB##. If not existing, take parent field BBB##.
Control Section
Normalized Record Field | Source/Content | Note |
---|---|---|
Source ID
|
From data source definitions
|
|
Original Source ID
|
From data source definitions
|
|
Source Record-ID
|
From data source definitions
|
|
Record ID
|
Source ID + Source Record-ID
|
|
Additional Record-ID
|
001#1 $$a
|
MAB identification number; may differ from Aleph system number.
|
ILS API ID
|
Original Source ID + Source Record-ID
|
|
Source Type
|
Not in use.
|
|
Source Format
|
From data source definitions
|
|
SourceSystem
|
From data source definitions
|
Display Section
Display Element | Source | Note |
---|---|---|
Resource Type
|
050 ## Pos. 10
OR
051 ## Pos. 1
OR
052 ## Pos. 1
OR
051 ## Pos. 0
OR
052 ## Pos. 0
OR
050 ## Pos. 5
OR
334 ## $$a
|
Mapping rule 4.
The following mapping tables are used:
|
Title
|
331 ## $$a
333 ## $$a
335 ## $$a
360 ## $$a
089 #1 $$a
304 #1 $$a
310 #1 $$a
340 #1 $$a – 352#1 $$a
(every 4th tag)
341 #1 $$a
|
Mapping rule 2.
Multiple occurrences of 360 and 089 are merged and delimited by “_:_”
Fields are merged in the following way:
331#2_/_333#2_:_335#2._360#2._089#1
._331#1_/_333#1_:_335#1._360#1._304#1._310_=_340
|
Creator
|
100 #1 $$a
104 a1 $$a
108 a1 $$a
(every 4th tag, until 136_1)
200 #1 $$a
204 a1 $$a
208 a1 $$a
|
Mapping rule 1.
Transformation routines “Get author first name” and “Get author last name” are used to reverse the author’s names.
Multiple occurrences are merged and delimited by “;_” in the listed sequence of the fields:
100_1, 104a1, 108a1, ... ,136a1, 200_1, 204a1, 208a1
|
Contributor
|
100 #1 $$a, $$b
Take any indicator except “_”
104 #1 $$a - 196#1 $$a, $$b
(every 4th tag)
Take any indicator except “a”
200 #1 $$a
Take any indicator except “_”
204 #1 $$a - 208#1 $$a (every 4th tag)
Take any indicator except “a”
|
Mapping rule 1.
Transformation routines “Get author first name” and “Get author last name” are used to reverse the author’s names.
Subfields $$a and $$b (relator term) are merged and delimited by blank.
|
Edition
|
403 ## $$a
407 ## $$a
|
Mapping rule 4.
Fields are merged and delimited by a semicolon:
403##;_407##
|
Publisher
|
410 ## $$a
412 ## $$a
415 ## $$a
417 ## $$a
418 ## $$a, $$g
|
Mapping rule 4.
Fields are merged and delimited by "_:_"
First place/publisher: 410##_:_412##
Second place/publisher: 415##_:_417##
Third and subsequent places/publisher: 418##
418: subfields $$a and $$g are merged and delimited by "_:_"
|
Creation date
|
405 #1 $$a
OR
425 b1 $$a, 425 c1 $$a
OR
425 #1 $$a
OR
425 a1 $$a
OR
425 p1 $$a
|
Mapping rule 1.
425b1 and 425c1 are merged. Transformation routine “Add to end of string” is used to add hyphen to 425b1.
Example: 1980_-_1985
|
Physical format
|
433 #1 $$a Take any indicator except "a".
434 #1 $$a
437 #1 $$a
653 #1 $$a
|
Mapping rule 1.
Fields are merged in the following way:
433#1_:_434_1_+_437#1._-_653#1
Multiple occurrences of 434_1 and 652#1 are merged and delimited by ",_"
|
Is Part Of
|
525 #1 $$p, $$a
590 #1 $$p, $$a - 599 #1 $$p, $$a
Take any indicator except blank and "s".
|
Mapping rule 1.
Subfields $$p and $$a of field 525#1 are merged and delimited by ":_"
Fields are merged like this:
525#1._-_In:_590#1_/_591#1._592#1._-_593#1._-_594#1,_595#1
._-_596#1._-_(597#1)._-_598#1._-_599#1
Multiple occurrences of 525#1 and 599#1 are merged and delimited by "._-_"
|
Identifier
|
540 #1 $$a, $$b
541 #1 $$a, $$b
542 #1 $$a, $$b
543 #1 $$a, $$b
Take 54X with any indicator except “z”.
|
Mapping rule 1.
Transformation routine “Add to beginning of string” is used to prefix fields with following values:
540: “ISBN ”
541: “ISMN ”
542: “ISSN ”
543: “ISRN ”
Subfields $$a and $$b are merged and delimited by “_:_”
Multiple occurrences are merged and delimited by “_;_” in the listed sequence of the fields: 540#1, 541#1, 542#1, 543#1
|
Subject
|
902 ## $$a - 947 ## $$a (every 5th tag)
Take any subfields except $$1 and $$9
710 ## $$a
711 ## $$a
720 ## $$a
740 ## $$a
|
Mapping rule 2.
All repeatable fields of 902## with all subfields are considered as one subject (the same for 907##ff - 947##). Fields and subfields are merged and delimited by "_/_".
|
Description
|
517 #1 $$p, $$a
524 #1 $$p, $$a
750 #1 $$a
753 #1 $$a
756 #1 $$a
|
Mapping rule 1.
Subfields $$p and $$a of field 5XX#1 are merged and delimited by ":_"
|
Language
|
037 b# $$a
OR
037 a# $$a
OR
037 c# $$a
|
Mapping rule 3.
The language is stored in ISO 639-2.
Multiple occurrences of subfields $$a of 037 ## are merged with ";_"
|
Relation
|
451#1 $$a - 491#1 $$a
(every 10th tag)
522 #1 $$p, $$a
526 #1 $$p, $$a - 533 #1 $$p, $$a
534 #1
|
Mapping rule 1.
Subfields $$p and $$a of field 52X#1 are merged and delimited by ":_"
Multiple occurrences are merged and delimited by "._-_" in the listed sequence of the fields: 451#1, 461#1, 471#1, 481#1, 491#1, 522#1, 526#1, 527#1, 528#1, 529#1, 530#1, 531#1, 532#1, 533#1, 534#1
|
Source
|
From PNX:
control/sourceid
|
|
Coverage
|
---
|
|
Rights
|
---
|
|
Library level availability
|
AVA, created by the Aleph expand routine "expand_doc_bib_avail
|
The Availability subfields are created as follows:
$$I Institution – based on institution look-up table based on AVA$$a
$$L Library – based on library look-up table based on AVA$$b
$$1 Sublocation – from AVA$$c
$$2 Call number – from AVA$$d
$$S Availability status – based on AVA$$e
$$3 No. of items – from AVA$$f
$$4 No. of unavailable items – from AVA$$g
$$5 multi-volume flag – from AVA$$h
$$6 number of loans – from AVA $$i
$$9 – For more information, refer to section Adding $$9ONLINE to Library Level Availability.
$$P location priority – from AVA $$p
$$X source institution code (Aleph ADM) – from AVA $$a
$$Y source library code (Aleph sublibrary) – from AVA $$b
$$Z source sublocation code (Aleph collection) – from AVA $$j
|
Vernacular title
|
---
|
|
Uniform title
|
304 #1 $$a
310 #1 $$a
|
Mapping rule 1.
Fields are merged and delimited by semicolon:
304 #1_;_310 #1
|
Links
Type of Link | Source | Note |
---|---|---|
OpenURL
|
Based on resource type from display:
If type=article then:
$$Topenurl_article
Otherwise:
$$Topenurl_journal
|
SFX has two sources for Primo: one for articles, in which case the data is used, and one for journals, in which case the date is ignored. There is a different template per source.
|
Backlink
|
$$Taleph_backlink
|
|
Link to Resource
|
655 #1 $$u, $$z
655 #1 $$g, $$z
552 #1 $$a
|
Mapping rule 1.
655: If $$z exists, take text from this subfield and add it to $$D as description for link. If $$z doesn't exist, add description "Link to Resource" to $$D.
|
Thumbnail
|
$$Tamazon_thumb
OR
$$Tsyndetics_thumb
OR
$$Tgoogle_thumb
|
|
Link to TOC
|
---
|
|
Link to Abstract
|
---
|
|
OpenURL_fulltext
|
Based on resource type from display:
If type=article then:
$$Topenurl_article
Otherwise:
$$Topenurl_journal
|
SFX has two sources for Primo: one for articles, in which case the data is used, and one for journals, in which case the date is ignored. There is a different template per source.
|
OpenURL_servicetext
|
---
|
|
Link to Holdings
|
$$Taleph_holdings
|
|
Link to Request
|
---
|
|
Link to Review
|
---
|
|
Link to Price
|
---
|
|
Additional links
|
---
|
|
Link to Holdings_Available
|
---
|
|
Link to Holdings_Unavailable
|
---
|
|
Link to Holdings_Doesnotexist
|
---
|
|
Link to UC
|
---
|
|
Link to Finding Aid
|
---
|
|
Link to Excerpt
|
---
|
Search
Index | Source tag | Notes |
---|---|---|
Creator/Contributor
|
100 ## - 196 ## $$a (every 4th tag)
101 ## - 197 ## $$a (every 4th tag)
800 ## $$a, 801 ## $$a
806 ## $$a, 808 ## $$a
812 ## $$a, 813 ## $$a
818 ## $$a, 819 ## $$a
824 ## $$a, 825 ## $$a
200 ## - 296 ## $$a (every 4th tag)
201 ## - 297 ## $$a (every 4th tag)
802 ## $$a, 803 ## $$a
808 ## $$a, 809 ## $$a
814 ## $$a, 815 ## $$a
820 ## $$a, 821 ## $$a
826 ## $$a, 827 ## $$a
|
Mapping rule 2.
The headings form of persons (not the cross references) are considered twice:
Copy as is
Transformation routine "Normalize author" is used to take the last name + the initial character from the first name
|
Title
|
089 #1 $$a
304 ## $$a
310 ## $$a
331 ## $$a
333 ## $$a
335 ## $$a
360 ## $$a
|
Mapping rule 2.
|
Description
|
517 #1 $$a
524 #1 $$a
750 #1 $$a
753 #1 $$a
756 #1 $$a
|
Mapping rule 1.
|
Subject
|
902## $$a - 947## $a (every 5th tag)
Take any subfield except $$1 and $$9
710 ## $$a
711 ## $$a, $$v
720 ## $$#
740 ## $$#
|
Mapping rule 2.
|
Full Text
|
---
|
|
General
|
400 ## $$a
403 ## $$a
407 ## $$a
410 ## $$a
412 ## $$a
415 ## $$a
417 ## $$a
418 ## $$a
519 ## $$a
|
Mapping rule 4.
|
Source ID
|
From PNX:
control/sourceid
|
|
Record ID
|
From PNX:
control/recordid
|
|
ISBN
|
540 ## $$a
|
Mapping rule 4.
|
ISSN
|
541 ## $$a
|
Mapping rule 4.
|
TOC
|
---
|
|
Resource type
|
From PNX:
display/type
|
|
Format
|
---
|
|
Creation date
|
425 a1 $$a OR 425p1 $$a OR
425 b1 $$a OR 425c1 $$a OR
425 #1 $$a OR 595_1 $$a OR
425a 2 $$a OR 425p2 $$a OR
425 b2 $$a OR 425c2 $$a OR
425 #2 $$a OR 595_2 $$a OR
089 #1 $$a
ADD
619 a1 $$a OR 619 b1 $$a OR
619 #1 $$a
ADD
674 #1 $$a
|
Mapping rule 3.
|
Record type
|
---
|
|
Additional Title
|
340 #1 $$a - 355 #1 $$a
361 #1 $$a
365 #1 $$a
370 #1 $$a
376 #1 $$a
451 #1 $$a - 491 #1 $$a (every 10th tag)
454 #1 $$a - 494 #1 $$a (every 10th tag)
621 #1 $$a, 624 #1 $$a, 627 #1 $$a, 630 #1 $$a
670 #1 $$a
675 #1 $$a
804 #1 $$a, 805 #1 $$a
810 #1 $$a, 811 #1 $$a
816 #1 $$a, 817 #1 $$a
822 #1 $$a, 823 #1 $$a
828 #1 $$a, 829 #1 $$a
|
Mapping Rule 1.
|
Additional resource record ID
|
From PNX:
control/addsrcrecordid
|
|
Search Scope
|
From PNX:
delivery/institution
control/sourceid (for example the data source is added as a scope)
|
|
Restricted Search Scope
|
---
|
|
Scope
|
Copies from the Search scope and Restricted search scope from the sections above
|
|
Syndetics full text
|
---
|
|
Syndetics TOC
|
---
|
|
Alternative title
|
---
|
Sort
Sort type | Source/Content | Note |
---|---|---|
Creation date
|
425 a1 $$a OR 425 p1 $$a OR
425 b1 $$a OR 425 c1 $$a OR
425 #1 $$a OR 595 #1 $$a OR
425 a2 $$a OR 425 p2 $$a OR
425 b2 $$a OR 425 c2 $$a OR
425 #2 $$a OR 595 #2 $$a OR
089 #1 $$a
|
Mapping rule 1.
Analyze the content of the field and use just the four numeric characters. If this shouldn't be possible (because of the content of the field) or if a source year field is missing use "0000" as default year.
|
Facets
Facet | Source | Note |
---|---|---|
Language
|
037 b# $$a
OR
037 a# $$a
OR
037 c# $$a
|
Mapping rule 3.
|
Creation date
|
425 a1 $$a OR 425 p1 $$a OR
425 b1 $$a OR 425 c1 $$a OR
425 #1 $$a OR 595 #1 $$a OR
425 a2 $$a OR 425 p2 $$a OR
425 b2 $$a OR 425 c2 $$a OR
425 #2 $$a OR 595 #2 $$a OR
089 #1 $$a
ADD
619 a1 $$a OR 619 b1 $$a OR
619 #1 $$a
ADD
674 #1 $$a
|
Mapping rule 1.
Analyze the content of the field and use just the four numeric characters. If this shouldn't be possible (because of the content of the field) or if a source year field is missing use "0000" as default year.
|
Topic
|
902 ## - 947 ## $$# (every 5th tag)
Any subfields except $$1 and $$9
710 ## $$a
711 ## $$a
740 ## $$a
|
Mapping rule 2.
All subfields of 902## are considered as one topic (the same for 907##ff - 947##). Subfields are merged with "_/_".
For example, 902#1$$p_/_$$t
|
Collection
|
AVA $$b
|
Mapping table "ILS library codes" is used.
|
Top-level
|
334 #1 $$a
OR
655 #1 $$u or $$g
OR
552 #1 $$a
OR
051 Pos. 0
OR
052 Pos. 0
OR
Constant=Book
|
The following mapping tables are used:
"toplevel_334"
"type_051_0"
"type_052_0"
ELSE: Constant "Book"
|
Pre-filter
|
Based on the field Type of the display section a 1:1 mapping should be done:
Book -> Books
Journal -> Journals
Article -> Articles
Text Resource -> Books
Audio -> Audio-Video
Video -> Audio-Video
Image -> Images
Map -> Maps
Score -> Scores
|
|
Resource type
|
Create this based on the Resource type field from display section as follows.
Book ' books
Journal ' journals
Article ' articles
Text Resource ' books
Audio ' audio-video
Video ' audio-video
Image ' images
Map ' maps
Score ' scores
Other ' other
|
In some cases, two values should be created, each as separate field.
|
Creator/Contributor
|
100 ## $$a - 196 ## $$a (every 4th tag)
800 ## $$a - 824 ## $$a (every 6th tag)
200 ## $$a - 296 ## $$a (every 4th tag)
802 ## $$a - 826 ## $$a (every 6th tag)
|
Mapping rule 2.
Transformation routine "Normalize author" is used.
|
Physical format
|
---
|
|
File size
|
---
|
|
Related record
|
331 #2 $$a
451 #1 $$a
461 #1 $$a
471 #1 $$a
481 #1 $$a
491 #1 $$a
|
451 - 491: Transformation routine "Take until first occurrence of" is used to take only characters until the first semicolon appears in the fields.
|
Genre
|
---
|
|
Language
|
037 b# $$a
OR
037 a# $$a
OR
037 c# $$a
|
Mapping rule 3.
|
Creation date
|
425 a1 $$a OR 425p1 $$a OR
425 b1 $$a OR 425c1 $$a OR
425 #1 $$a OR 595_1 $$a OR
425 a2 $$a OR 425p2 $$a OR
425 b2 $$a OR 425c2 $$a OR
425 #2 $$a OR 595_2 $$a OR
089 #1 $$a
ADD
619 a1 $$a OR 619 b1 $$a OR
619 #1 $$a
ADD
674 #1 $$a
|
Mapping rule 1.
Analyze the content of the field and use just the four numeric characters. If this shouldn't be possible (because of the content of the field) or if a source year field is missing use "0000" as default year.
|
Topic
|
902 ## - 947 ## $$# (every 5th tag)
Any subfields except $$1 and $$9
710 ## $$a
711 ## $$a
740 ## $$a
|
Mapping rule 2.
All subfields of 902## are considered as one topic (the same for 907##ff - 947##). Subfields are merged with "_/_".
For example, 902#1$$p_/_$$t
|
Collection
|
AVA $$b
|
Mapping table "ILS library codes" is used.
|
Top-level
|
334 #1 $$a
OR
655 #1 $$u or $$g
OR
552 #1 $$a
OR
051 Pos. 0
OR
052 Pos. 0
OR
Constant=Book
|
The following mapping tables are used:
"toplevel_334"
"type_051_0"
"type_052_0"
ELSE: Constant "Book"
|
Pre-filter
|
Based on the field Type of the display section a 1:1 mapping should be done:
Book -> Books
Journal -> Journals
Article -> Articles
Text Resource -> Books
Audio -> Audio-Video
Video -> Audio-Video
Image -> Images
Map -> Maps
Score -> Scores
|
|
Resource type
|
Create this based on the Resource type field from display section as follows:
Book -> books
Journal -> journals
Article -> articles
Text Resource -> books
Audio -> audio-video
Video -> audio-video
Image -> images
Map -> maps
Score -> scores
Other -> other
|
In some cases, two values should be created, each as separate field.
|
Creator/Contributor
|
100 ## $$a - 196 ## $$a (every 4th tag)
800 ## $$a - 824 ## $$a (every 6th tag)
200 ## $$a - 296 ## $$a (every 4th tag)
802 ## $$a - 826 ## $$a (every 6th tag)
|
Mapping rule 2.
Transformation routine "Normalize author" is used.
|
Physical format
|
---
|
|
File size
|
---
|
|
Related record
|
331 #2 $$a
451 #1 $$a
461 #1 $$a
471 #1 $$a
481 #1 $$a
491 #1 $$a
|
451 - 491: Transformation routine "Take until first occurrence of" is used to take only characters until the first semicolon appears in the fields.
|
Genre
|
---
|
Duplicate Record Detection Vector
Currently, two types of record matching vectors exist:
-
T1 – for non-serials
-
T2 – for serials
Vector for T1 - "non-serials"
Field ID | Nature of field | Content of Field/Source Tag + Subfield | Note |
---|---|---|---|
T
|
Type
|
FMT
052 #1
|
Mapping rule 1.
The type value "T" will be determined based on tag FMT and tag 052:
If tag FMT is "MH" ADD tag 052 is present then the type is T2
All other records with tag FMT "MH" are type T1 (for example, tag 052 is missing)
Records with tag FMT "MU" are not part of the duplication detection.
|
The following fields are for the candidate selection:
|
|||
C1
|
UnivID, NBN, Univ_ID_Invalid
|
574 #1
|
Mapping rule 1.
|
C2
|
ISXN, ISXN_invalid
|
540 #1 $$a
(any indicator except z)
541 #1 $$a
543 #1 $$a
634 #1 $$a
(any indicator except z)
|
Mapping rule 1.
Transformation routine "Drop string" is used to remove hyphen.
Multiple occurrences are merged and delimited by "_;_".
|
C3
|
Short title+author
|
331 #1 $$a
ADD
100 #1 $$a
OR
200 #1 $$a
|
Mapping rule 1.
Transformation routine "Take string" is used to normalize only the first 25 characters of title.
Transformation routine "Take string" is used to normalize only the first 10 characters of title.
|
C4
|
Year
|
425 a1 $$a
OR
425 p1 $$a
OR
425 #1 $$a
OR
595 #1 $$a
|
Mapping rule 1.
Only the four numeric characters are used. If it's not possible to identify the year (because of the content of the field) or if a year source field is missing, "0000" is used as the default year.
|
C5 - C10
|
---
|
---
|
|
The following fields are for the matching program:
|
|||
F1
|
UnivID NBN
|
Use C1
|
|
F2
|
Univ_Invalid
|
---
|
|
F3
|
ISXN
|
540 a1 $$a
541 a1 $$a
543 a1 $$a
634 a1 $$a
|
Mapping rule 1.
Multiple occurrences are delimited by a semicolon.
Transformation routine "Drop string" is used to remove the hyphen.
|
F4
|
ISXN_Invalid
|
540 #1 $$a
(only indicator blank and b)
541 #1 $$a
543 #1 $$a
634 #1 $$a
(only indicator blank and b)
|
Mapping rule 1.
Multiple occurrences are delimited by a semicolon.
Transformation routine "Drop string" is used to remove the hyphen.
|
F5
|
Short title+author
|
C3 of the candidate section
|
|
F6
|
Year
|
C4 of the candidate section
|
|
F7
|
Full title
|
331 #1 $$a
|
Mapping rule 1.
|
F8
|
Place of publication
|
410 #1 $$a
|
Mapping rule 1.
Transformation routine "Take string" is used to normalize only the first 5 normalized characters of the field.
|
F9
|
Pagination
|
433 #1 $$a
|
Mapping rule 1.
|
F10
|
Publisher
|
412 #1 $$a
|
Mapping rule 1.
Transformation routine "Take string" is used to normalize only the first 5 normalized characters of the field.
|
F11
|
Main entry (author, corporate body, meeting)
|
100 #1 $$a
OR
200 #1 $$a
OR
100 #1 $$a
|
Mapping rule 1.
|
F12
|
Physical format
|
334 #1 $$a
|
|
Mapping rule 1.
|
|||
F13
|
Edition
|
403 #1
OR
400 #1
|
Mapping rule 1.
Transformation routine "Take string" is used to normalize only the first 5 normalized characters of the field.
|
Vector for T2 - "serials"
Field ID | Nature of field | Content of Field/Source Tag + Subfield | Note |
---|---|---|---|
T
|
Type
|
FMT
052 #1
|
Mapping rule 1
The type value T2 will be determined based on tag FMT and tag 052:
If tag FMT is "MH" AND tag 052 is present then the type is T2
All other records with tag FMT "MH" are type T1 (i.e. tag 052 is missing)
Records with tag FMT "MU" are not part of the duplication detection.
|
The following fields are for the candidate selection:
|
|||
C1
|
Univ_ID, ZDB-ID, UnivID_invalid
|
025 z1 (= ZDB ID)
|
Mapping rule 1.
Rule for the contents of the field:
If the field contains only alphabetic characters ignore the content for the field UnivID.
|
C2
|
ISSN, ISSN_invalid, ISSN_Cancelled
|
542 #1 $$a
(any indicator except z)
635 #1 $$a
|
Mapping rule 1.
Transformation routine "Drop string" is used to remove hyphen.
Multiple occurrences are merged and delimited by "_;_".
|
C3
|
Short Title + corporate body
|
310 #1 $$a
OR
331 #1 $$a
ADD
200 #1 $$a
|
Mapping rule 1.
Transformation routine "Take string" is used to normalize only the first 25 characters of title.
Transformation routine "Take string" is used to normalize only the first 10 characters of title.
|
C4
|
Place of publication
|
410 #1 $$a
|
Mapping rule 1.
Transformation routine "Take string" is used to normalize only the first 5 characters.
|
C5 - C10
|
---
|
---
|
Not in use.
|
The following fields are for the matching program:
|
|||
F1
|
ZDB-ID
|
Use C1
|
|
F2
|
Univ_invalid
|
---
|
|
F3
|
ISSN
|
542 a1 $$a
635 a1 $$a
|
Mapping rule 1.
Multiple occurrences are delimited by semicolon.
Transformation routine "Drop string" is used to remove hyphen.
|
F4
|
ISSN_Invalid
|
542 #1 $$a (R
635 #1 $$a (R)
|
Mapping rule 1.
Multiple occurrences are delimited by semicolon.
Transformation routine "Drop string" is used to remove hyphen.
|
F5
|
ISSN cancelled
|
-
|
|
F6
|
Year of first volume
|
425 b1
|
Mapping rule 1.
Analyze the content of the field and use just the four numeric characters. If this shouldn't be possible (because of the content of the field) or if a year source field is missing use "0000" as default year.
|
F7
|
Full title
|
310 #1
OR
331 #1
|
Mapping rule 1.
|
F8
|
Short title + corporate body
|
take C3 of the candidate section
|
|
F9
|
Country of publication
|
---
|
|
F10
|
Place of publication
|
410 #1 $$a
|
Mapping rule 1.
Transformation routine "Take string" is used to take only the first 5 normalized characters of the field.
|
F11
|
Main entry ( corporate body, meeting)
|
200 #1 $$a
|
Mapping rule 1.
|
F12
|
Physical format
|
334 #1 $$a
|
Mapping rule 1.
|
F13
|
Publisher
|
412 #1 $$a
|
Mapping rule 1.
Transformation routine "Take string" is used to take only the first 5 normalized characters of the field.
|
FRBRization
Field ID | Source (value of $$K for K fields) | Key part type (value of $$A for K fields) | Note |
---|---|---|---|
T
|
Always 1
|
||
K1
|
100 #1 $$a
OR
200 #1 $$a
OR
800 ## $$a
OR
200 ## $$a
|
A
|
Following transformation routines are used:
"Delete characters", parameter: |[]'
"Replace characters", parameter: ,$~'^%*/\?@.:;<>{}}-()"!¿¡@@^
"Character conversion" using the NACO_diacritics character conversion table.
"Lower case"
|
K2
|
304 #1 $$a
|
TO
|
Following transformation routines are used:
"Delete characters", parameter: |[]'
"Replace characters", parameter: ,$~'^%*/\?@.:;<>{}}-()"!¿¡@@^
"Character conversion" using the NACO_diacritics character conversion table.
"Lower case"
|
K3
|
From PNX:
display/title
|
T
|
|
K4 - K10
|
---
|
---
|
Delivery and Scoping
Delivery Field | Source | Additional Normalization Notes |
---|---|---|
Institution
|
AVA $$a
OR
Constant "MAB
|
Mapping rule 1.
Mapping table "ILS Institution Codes" is used.
|
Delivery category
|
334 #1 $$a
655 #1 $$u, $$g
522 b1 $$a
050 Pos. 3
|
Mapping rule 4.
The following mapping tables are uses:
334 : "delcategory_334"
050 : "delcategory_050_3"
If no value can be assigned according to the 4 fields, the constant "Physical item" should be used.
|
Restricted delivery scope
|
---
|
Customer specific; if needed to be defined during implementation phase.
|
Ranking
Local mapping required as relevant.
Booster Field | Source | Additional Normalization Notes |
---|---|---|
booster1
|
1 or as added by enrichment program
|
|
booster2
|
Not in use.
|
Enrichment
Local mapping required as relevant.
Enrichment Field | Source | Additional Normalization Notes |
---|---|---|
Classification LCC
[classificationlcc]
|
---
|
|
Classification DDC
[classificationddc]
|
---
|
|
Classification UDC
|
---
|
|
Classification RVK
|
700 g# $$a
(= RVK classification code as source for the enrichment process)
|
Mapping rule 2.
Multiple occurrences are not concatenated.
|
Fulltext
|
---
|
|
TOC
|
---
|
|
Abstract
|
---
|
|
Review
|
---
|
|
Rank parent/child
|
FMT
|
IF FMT= "MU" then set constant "mab-u"
|
Rank number of copies
|
---
|
|
Rank date first copy
|
---
|
|
Rank number of loans
|
---
|
|
Availability
|
---
|
Additional Data
This includes multiple occurrences in separate fields.
Additional Data Field | Source | Additional Normalization Notes |
---|---|---|
Author Last Name
|
100 ## $$a - 196 ## $$a (every 4th tag)
101 ## $$a - 197 ## $$a (every 4th tag)
|
Mapping rule 2.
Transformation routine "Get author last name" is used.
|
Author First Name
|
100 ## $$a - 196 ## $$a (every 4th tag)
101 ## $$a - 197 ## $$a (every 4th tag)
|
Mapping rule 2.
Transformation routine "Get author first name" is used.
|
Author Initial
|
---
|
|
Author Initial 1 (first initial)
|
---
|
|
Author Initial Middle
|
---
|
|
Author suffix
|
---
|
|
Author
|
100 ## $$a - 196 ## $$a (every 4th tag)
101 ## $$a - 197 ## $$a (every 4th tag)
|
Mapping rule 2.
|
Corporate author
|
200 ## $$a - 296 ## $$a (every 4th tag)
201 ## $$a - 297 ## $$a (every 4th tag)
802 ## $$a - 82 6## $$a (every 6th tag)
803 ## $$a - 827 ## $$a (every 6th tag)
|
Mapping rule 2.
|
Additional author
|
800 ## $$a - 824 ## $$a (every 6th tag)
801 ## $$a - 825 ## $$a (every 6th tag)
|
Mapping rule 2.
|
Series author
|
---
|
|
Book title
|
From PNX:
display/title
|
Only if PNX: display/type not article OR journal
|
Article title
|
From PNX:
display/title
|
Only if PNX: display/type = article
|
Journal title
|
From PNX:
display/title
|
Only if PNX: display/type = journal
|
Short title
|
331 ## $$a
335 ## $$a
360 ## $$a
089 #1 $$a
|
Mapping rule 2.
Transformation routine "Take substring / 0@@25" is used to normalize only the first 25 characters.
|
Additional title
|
340 ## $$a - 355 ## $$a
361 ## $$a
365 ## $$a
370 ## $$a
376 ## $$a
451 ## $$a - 491 ## $$a
(every 10th tag)
454 ## $$a - 494 ## $$a
(every 10th tag)
503 ## $$a
621 ## $$a, 624 ## $$a, 627 ## $$a, 630 ## $$a
670 ## $$a
675 ## $$a
804 ## $$a - 828 ## $$a (every 6th tag)
805 ## $$a - 829 ## $$a (every 6th tag)
|
Mapping Rule 2.
|
Series title
|
---
|
|
Date
|
425 a1 $$a OR 425 p1 $$a OR
425 b1 $$a OR 425 c1 $$a OR
425 #1 $$a OR 595 #1 $$a OR
425 a2 $$a OR 425 p2 $$a OR
425 b2 $$a OR 425 c2 $$a OR
425 #2 $$a OR 595 #2 $$a OR
089 #1 $$a
ADD
619 a1 $$a OR 61 9b1 $$a OR
619 #1 $$a
ADD
674 #1 $$a (R)
|
Mapping rule 4.
|
RISDate
|
---
|
|
Additional date
|
---
|
|
Volume
|
---
|
|
Issue
|
---
|
|
Part
|
---
|
|
Season
|
---
|
|
Quarter
|
---
|
|
Start page
|
---
|
|
End page
|
---
|
|
Pages
|
---
|
|
Article number
|
553 #1 $$a
|
Mapping rule 1.
|
ISSN
|
542 #1 $$a
|
Mapping rule 1.
|
eISSN
|
||
ISBN
|
540 #1 $$a
|
Mapping rule 1.
|
CODEN
|
376 b1 $$a
|
Mapping rule 1.
|
SICI
|
---
|
|
Metadata format
|
IF 519 exists -> dissertation
Else based on Resource type from display:
|
Mapping table "Metadata format" is used.
|
Genre
|
The Genre mapping table maps the resource type from the display section of the PNX to the genre that is required by the OpenURL.
|
Use Genre mapping table.
|
RISType
|
---
|
|
Notes
|
519 #1 $$a
517 #1 $$p, $$a
|
Mapping rule 1.
Subfields $$p and $$a of fields 5XX#1 are merged like this:
$$p + suffix ":_" is the prefix for $$a
|
Abstract
|
750 #1 $$a
753 #1 $$a
756 #1 $$a
|
Mapping rule 1.
|
City of Publication
|
410 ## $$a
415 ## $$a
|
Mapping rule 4.
|
Publisher
|
412 ## $$a
417 ## $$a
|
Mapping rule 4.
|
Miscellaneous 1
|
---
|
|
Miscellaneous 2
|
---
|
|
Miscellaneous 3
|
---
|
|
OCLC ID
|
---
|
|
DOI
|
552 #1 $$a
|
Mapping rule 1.
|
URL
|
655 #1 $$u
655 #1 $$g
552 #1 $$a
|
Mapping rule 1.
|
Browse
The system can create multiple occurrences in separate fields.
Browse field | Source | Additional normalization notes |
---|---|---|
Institution
|
PNX: delivery/institution
|
|
Author
|
100 #1 $$a
104 a1 $$a
108 a1 $$a
(every 4th tag, until 136_1)
200 #1 $$a
204 a1 $$a
208 a1 $$a
|
$$D (display form) and $$E (normalized form) are created.
|
Title
|
331 ## $$a
335 ## $$a
360 ## $$a
089 #1 $$a
304 #1 $$a
310 #1 $$a
340 #1 $$a - 352#1 $$a
(every 4th tag)
341 #1 $$a - 353#1$$a
(every 4th tag)
|
Fields are merged in the following way for $$D (display form):
331#2 : 335#2
331#2. 089#1
331#1 : 335#1
Fields are merged in the following way for $$E (normalized form):
331#2 335#2
331#2 089#1
331#1 335#1
|
Subject
|
902 ##
(Take any subfields except $$1 and $$9)
907 ##
(Take any subfields except $$1,$$9,$$a)
912 #1
(every 5th tag until 947; take any subfields except $$1,$$9,$$a)
710 ## $$a
711 ## $$a
740 ## $$a
|
$$D (display form) and $$E (normalized form) are created.
|
Call number
|
AVA $$d (for call number)
AVA $$a (for institution)
|
$$D (display form) and $$E (normalized form) are created.
|