Rick Crume, a genealogy consultant and the article’s author, says IPTC metadata “can be extremely useful for savvy archivists […] IPTC standards can help future-proof your metadata. That data becomes part of the digital photo, contained inside the file and preserved for future software programs.”
Crume quotes Ken Watson from All About Digital Photos saying “[IPTC] is an internationally recognized standard, so your IPTC/XMP data will be viewable by someone 50 or 100 years from now. The same cannot be said for programs that use some proprietary labelling schemes.”
Crume then adds: “To put it another way: If you use photo software that abides by the IPTC/XMP standard, your labels and descriptive tags (keywords) should be readable by other programs that also follow the standard. For a list of photo software that supports IPTC Photo Metadata, visit the IPTC’s website.“
“[IPTC] is an internationally recognized standard, so your IPTC/XMP data will be viewable by someone 50 or 100 years from now”
The article goes on to recommend particular software choices based on IPTC’s list of photo software that supports IPTC Photo Metadata. In particular, Crume recommends that users don’t switch from Picasa to Google Photos, because Google Photos does not support IPTC Photo Metadata in the same way. Instead, he recommends that users stick with Picasa for as long as possible, and then choose another photo management tool from the supported software list.
Similarly, Crume recommends that users should not move from Windows Photo Gallery to the Windows 10 Photos app, because the Photos app does not support IPTC embedded metadata.
Crume then goes on to investigate popular genealogy sites to examine their support for embedded metadata, something that we do not cover in our photo metadata support surveys.
Following on with our quarterly update cycle, the IPTC NewsCodes Working Group has released the Q2 2022 update of IPTC NewsCodes, including updates to the Media Topic, Subject Code, and Digital Source Type vocabularies.
Media Topic updates
- Translation changes:
- A new language translation for “New Norwegian” (Norwegian nynorsk, no-NN) has been added to all labels. The existing Norwegian labels previously tagged with “no” are now tagged as “no-NB” for Norwegian bokmål. Thanks very much to NTB for providing the update.
- Label and definition changes:
- medtop:20000446 diseases and conditions
- medtop:20001230 corporate social responsibility -> environmental, social and governance policy (ESG)
- medtop:20000449 epidemic -> epidemic and pandemic
- medtop:20000451 virus disease -> viral disease
- medtop:20000452 AIDS -> HIV and AIDS
- medtop:20000457 medical conditions -> medical condition
- medtop:20000458 mental health and disorder
- medtop:20000463 health organisation
- medtop:20000464 health treatment -> health treatment and procedure
- medtop:20000466 dietary supplement
- medtop:20000467 medical drugs -> non-prescription drug
- medtop:20000468 prescription drugs -> prescription drug
- medtop:20000469 medical procedure/test -> medical test
- medtop:20000470 medicine -> health care approach
- medtop:20000474 western medicine -> conventional medicine
- medtop:20000480 government health care
- medtop:20001225 ophthalmology -> eye care
- Definition changes:
- medtop:07000000 health
- medtop:20000784 family planning
- medtop:20000454 heart disease
- medtop:20000456 injury
- medtop:20000461 health facility
- medtop:20000465 diet
- medtop:20001219 drug rehabilitation
- medtop:20001221 emergency care
- medtop:20000471 herbal medicine
- medtop:20000472 holistic medicine
- medtop:20000473 traditional Chinese medicine
- medtop:20000479 healthcare policy
- medtop:20000483 health insurance
- medtop:20000484 private health care
- medtop:20000486 medical service
- medtop:20000490 paediatrics
- Hierarchy moves:
- medtop:20000500 animal moves to become a child of medtop:20000441 nature
- medtop:20000507 flowers and plants moves to become a child of medtop:20000441 nature
- medtop:20001318 pests moves to become a child of medtop:20000500 animal
- medtop:20000494 animal disease moves to become a child of medtop:20000500 animal
- medtop:20000495 plant disease moves to become a child of medtop:20000507 flowers and plants
- medtop:20000460 obesity becomes a child of medtop:20000457 medical condition
- medtop:20000477 vaccine becomes a child of medtop:20000464 health treatment and procedure
- New terms:
- Retired terms:
- medtop:20001218 pandemic (use the new “epidemic and pandemic” term instead)
- medtop:20000450 plague (disease)
- medtop:20000453 retrovirus
- medtop:20000455 illness
- medtop:20000475 physical fitness
- medtop:20000476 preventative medicine
- medtop:20001220 general practice
- medtop:20000488 geriatric medicine
- medtop:20000489 obstetrics/gynaecology
- medtop:20001223 oncology
- medtop:20001222 orthopaedics
- medtop:20000713 pharmacology
- medtop:20001227 psychiatry
- medtop:20001224 radiology
- medtop:20000491 reproductive medicine
- medtop:20001226 surgical medicine
- medtop:20000493 non-human diseases
In a related tool update announcement, we have now added a handy “show retired terms” checkbox to the Media Topics interactive tree browser tool, and we default to only showing the active (non-retired) terms. The new option can be seen in the picture at the top of this article.
Digital Source Type vocabulary updates
After asking for feedback on a draft of the work a few months ago, we have updated the Digital Source Type vocabulary to support the emerging area of “Synthetic Media.”
The single term “softwareImage” has been retired, which means that while it is acceptable in legacy content, we no longer recommend its use. The term is now replaced with 9 new terms covering the spectrum from purely human creation through to purely machine image creation:
- Original media with minor human edits
- Composite of captured elements
- Algorithmically-enhanced media
- Data-driven media
- Digital art
- Virtual recording
- Composite including synthetic elements
- Trained algorithmic media
- Pure algorithmic media
- RETIRED: Created by software
To see more detail including the definition of each term, click the links above or view the entire IPTC Digital Source Type vocabulary.
Thanks to those both inside and outside of the IPTC community who gave feedback on our original proposal, your comments were very much appreciated.
Subject Code vocabulary updates – indicating its deprecated status
The IPTC Subject Code vocabulary was created over twenty years ago, in the year 2000. It was maintained through to 2010, but at that point the Media Topic vocabulary took over as IPTC’s preferred subject classification taxonomy. We will keep it on our vocabulary server, but we no longer recommend its use in projects due to some terms being out of date.
So we have put warnings on the pages of the Subject Code vocabulary that indicate its deprecated nature, and encourage users to look at Media Topic instead.
As always, the Media Topics vocabularies can be viewed in the following ways:
- In a collapsible tree view
- As a downloadable Excel spreadsheet
- On one page on the cv.iptc.org server
- In machine readable formats such as RDF/XML and Turtle using the SKOS vocabulary format: see the cv.iptc.org guidelines document for more detail.
For more information on IPTC NewsCodes in general, please see the IPTC NewsCodes Guidelines.
At the IPTC Spring Meeting in May 2022, IPTC’s Standards Committee voted to approve ninjs 1.4, the latest version in the 1.x track of IPTC’s standard for news content in the JSON format.
Johan Lindgren of TT Nyhetsbyrån, Lead of the IPTC News in JSON Working Group, said:
“After the launch of ninjs 2.0 in the autumn of 2021, we received requests to add some of the new 2.0 features to the first generation of ninjs, so that those who are using the 1.x branch of ninjs can use the new features without making breaking changes. So we are excited to publish version 1.4 of ninjs, where these features are included.”
Those changes include:
- New property contentcreated, denoting the date and time when the content of this ninjs object was originally created (as opposed to the date and time when the ninjs object itself was created). For example, an old photo that is now handled as a ninjs object may have a firstcreated and versioncreated of “2022-06-02T12:00:00+00:00”, but a contentcreated value of “1933-04-03T00:00:00+00:00”. The contents must be a valid JSON Schema date-time object.
- New property expires, showing “the date and time after which the Item is no longer considered editorially relevant by its provider.” Note that this is not the same as a rights-related expiration, it simply conveys the desire of the content creator to highlight the content until a certain time. A good example might be a football match preview, which would no longer be editorially relevant after the game commences. The contents must be a valid JSON Schema date-time object.
- New property rightsinfo, which holds an expression of rights to be applied to the content. It contains sub-properties langid (a URI which specifies the language used to specify rights such as RightsML or ODRL), and one of either linkedrights (containing a link to a remotely-hosted declaration of the rights associated with the content) or encodedrights (which includes an embedded encoding of the rights statements within the ninjs object).
Which version of ninjs should I choose for my project?
There might be some confusion since we have released ninjs 1.4 after the release of ninjs 2.0. Please note that this is simply an update to the 1.x branch of ninjs to make it easier for users who cannot upgrade to 2.x branch due to breaking changes.
If you are starting a new project that requires JSON-encoded news content, we recommend using ninjs 2.0. This version should be easiest for developers to work with.
If you are already using a 1.x version of ninjs, we recommend at least upgrading to version 1.4. This should be an easy change, because 1.4 is backwards-compatible with versions 1.0, 1.1, 1.2 and 1.3. We would also recommend upgrading to 2.0 if possible, but if not, 1.4 is the best version of the 1.x branch.
Supporting materials for ninjs 1.4 and ninjs 2.0 can be found at these locations:
- JSON Schema for ninjs 1.4: https://iptc.org/std/ninjs/ninjs-schema_1.4.json
- JSON Schema for ninjs 2.0: https://iptc.org/std/ninjs/ninjs-schema_2.0.json
- User Guide for ninjs (covering version 2.0 but mentioning 1.4): https://iptc.org/std/ninjs/userguide/
- The ninjs Generator has been updated to output 1.4-compatible JSON: https://www.iptc.org/std/ninjs/generator/
Thanks to Johan and the IPTC News in JSON Working Group for working on this release.
The IPTC took part in a panel on Diversity and Inclusion at the CEPIC Congress 2022, the picture industry’s annual get-together, held this year in Mallorca Spain.
Google’s Anna Dickson hosted the panel, which also included Debbie Grossman of Adobe Stock, Christina Vaughan of ImageSource and Cultura, and photographer Ayo Banton.
Unfortunately Abhi Chaudhuri of Google couldn’t attend due to Covid, but Anna presented his material on Google’s new work surfacing skin tone in Google image search results.
Brendan Quinn, IPTC Managing Director participated on behalf of the IPTC Photo Metadata Working Group, who put together the Photo Metadata Standard including the new properties covering accessibility for visually impaired people: Alt Text (Accessibility) and Extended Description (Accessibility).
Brendan also discussed IPTC’s other Photo Metadata properties concerning diversity, including the Additional Model Information which can include material on “ethnicity and other facets of the model(s) in a model-released image”, and the characteristics sub-property of the Person Shown in the Image with Details property which can be used to enter “a property or trait of the person by selecting a term from a Controlled Vocabulary.”
Some interesting conversations ensued around the difficulty of keeping diversity information up to date in an ever-changing world of diversity language, the pros and cons of using controlled vocabularies (pre-selected word lists) to cover diversity information, and the differences in covering identity and diversity information on a self-reported basis versus reporting by the photographer, photo agency or customer.
It’s a fascinating area and we hope to be able to support the photographic industry’s push forward with concrete work that can be implemented at all types of photographic organisations to make the benefits of photography accessible for as many people as possible, regardless of their cultural, racial, sexual or disability identity.
Where else can you hear about the difficulties of examining photo metadata in NFTs, see a lifelike image of a human being generated from pure data before your eyes, see how Wikidata can be used to take semantic fingerprints of news articles, and discover that an hour is nowhere near long enough to discuss simplifying machine-readable rights? Nowhere but the IPTC Meeting, of course! And this year’s Spring Meeting was the venue for all of this and much more.
We held the meeting virtually from Monday May 16 to Wednesday May 18th, and attending were over 70 people from at least 45 organisations across more than 20 countries.
Along with our usual Working Group updates and committee meetings, we invited speakers from several fascinating startups, services and projects at member companies. Here’s a quick summary of their sessions:
- We heard from Kairntech who are working on a classification system based on extracting entities from news stories and building a “semantic fingerprint” which can be used for cross-language classification, search and content enhancement
- The New York Times’ R&D Lab presented PaperTrail, a project to enhance the quality of the Times’ print archive through the use of machine learning to improve on basic OCR techniques (they’re looking for collaborators, more info coming soon!)
- Bria.ai showed us how an API can be used to enhance and create images and videos through the use of a custom GAN model trained in a “responsible AI” method
- Margaret Warren talked us through her efforts in creating and selling an NFT, looking at the process view the perspective of a photo metadata expert
- Consultant and author Henrik de Gyor talked us through the latest in synthetic media, which will be helpful in helping us to finalise our Digital Source Type vocabulary for synthetic media
- Laurent Le Meur from EDRLab presented his project’s recommendation on a Text and Data Mining Reservation Protocol, which can be used by publishers to restrict the rights of data miners in scraping any content for the purpose of analysis or building a model
- We heard from Dominic Young of Axate on his approach to offer pay-as-you-go payment options on paywalled news sites based on a simple pre-paid wallet mechanism.
We also had many announcements and discussions around IPTC standards, many of which we will be revealing in the coming months. One notable update is that the Standards Committee approved ninjs version 1.4 which we will release soon.
Thanks to all the IPTC members, Working Group leads, committee members and guests who made this member meeting one to remember.
The National Association of Broadcasters (NAB) Show wrapped up its first face-to-face event in three years last week in Las Vegas. In spite of the name, this is an internationally attended trade conference and exhibition showcasing equipment, software and services for film and video production, management and distribution. There were 52,000 attendees, down from a typical 90-100k, with some reduction in booth density; overall the show was reminiscent of pre-COVID days. A few members of IPTC met while there: Mark Milstein (vAIsual), Alison Sullivan (MGM Resorts), Phil Avner (Associated Press) and Pam Fisher (The Media Institute). Kudos to Phil for working, showcasing ENPS on the AP stand, while others walked the exhibition stands.
NAB is a long-running event and several large vendors have large ‘anchor’ booths. Some such as Panasonic and Adobe reduced their normal NAB booth size, while Blackmagic had their normal ‘city block’-sized presence, teeming with traffic. In some ways the reduced booth density was ideal for visitors: plenty of tables and chairs populated the open areas making more meeting and refreshment space available. The NAB exhibition is substantially more widely attended than the conference, and this year several theatres were provided on the show floor for sessions any ‘exhibits only’ attendee could watch. Some content is now available here: https://nabshow.com/2022/videos-on-demand/
For the most part this was a show of ‘consolidation’ rather than ‘innovation’. For example, exhibitors were enjoying welcoming their partners and customers face-to-face rather than launching significant new products. Codecs standardised during the past several years were finally reaching mainstream support, with AV1, VP9 and HEVC well-represented across vendors. SVT-AV1 (Scalable Vector Technology) was particularly prevalent, having been well optimised and made available to use license-free by the standard’s contributors. VVC (Versatile Video Coding), a more recent and more advanced standard, is still too computationally intensive for commercial use, though a small set made mention of it on their stands (e.g. Fraunhofer).
IP is now fairly ubiquitous within broadcast ecosystems. To consolidate further, an IP Showcase booth illustrating support across standards bodies and professional organisations championed more sophisticated adoption. A pyramid graphic showing a cascade of ‘widely available’ to ‘rarely available’ sub-systems encouraged deeper adoption.
Super Resolution – raising the game for video upscaling
One of the show floor sessions – “Improving Video Quality with AI” – presented advances by iSIZE and Intel. The Intel technology may be particularly interesting to IPTC members, and concerns “Super Resolution.” Having followed the subject for over 20 years, for me this was a personal highlight of the show.
Super Resolution is a technique for creating higher resolution content from smaller originals. For example, achieving a professional quality 1080p video from a 480p source, or scaling up a social media-sized image for feature use.
A few years ago a novel and highly effective new Super Resolution method was innovated (“RAISR”, see https://arxiv.org/abs/1606.01299); this represented a major discontinuity in the field, albeit with the usual mountain of investment and work needed to take the ‘R’ (research) to ‘D’ (development).
This is exactly what Intel have done, and the resulting toolsets will be made available at no cost at the company’s Open Visual Cloud repository at the end of May.
Intel invested four years in improving the AI/ML algorithms (having created a massive ground truth library for learning), optimising to CPUs for performance and parallelisation, and then engineering the ‘applied’ tools developers need for integration (e.g. Docker containers, FFmpeg and GStreamer plug-ins). Performance will now be commercially robust.
The visual results are astonishing, and could have a major impact on the commercial potential of photographic and film/video collections needing to reach much higher resolutions or even to repair ‘blurriness’.
Next year’s event is the centennial of the first NAB Show and takes place from April 15th-19th in Las Vegas.
– Pam Fisher – Lead, IPTC Video Metadata Working Group
With less than two weeks to go, we are pleased to announce the full agenda for the IPTC Spring Meeting 2022.
The IPTC Spring Meeting 2022 will be held virtually from Monday May 16th to Wednesday May 18th, from 1300 – 1800 UTC each day.
IPTC member representatives can view the full agenda and register at https://iptc.org/moz/events/spring-meeting-2022/
Highlights of the meeting include:
- Updates from all IPTC Working Groups, including Photo Metadata, Video Metadata, NewsCodes, Sports Content, NewsML-G2 and News in JSON
- Updates from the IPTC PR Committee and the IPTC Standards Committee, including votes on proposed new versions of IPTC standards
- Invited presentations from:
- United Robots, presenting their “robot journalism” system built for media companies
- Axate‘s micropayments system for publishers
- Kairntech presenting their content classification system used by Agence France-Presse among others
- Bria.ai‘s image generation and manipulation API backed with cutting-edge artificial intelligence
- Consultant Henrik de Gyor speaking on the latest developments in synthetic media
- Laurent Le Meur from EDRLab discussing the W3C Text and Data Mining Community Group’s recommendation for a Text and Data Mining Reservation Protocol
- Member presentations:
- Recently-joined IPTC members will have a chance to introduce themselves and their organisations to the IPTC membership
- The New York Times presenting their “Papertrail” system used to target advertising based on content metadata
- Margaret Warren from ImageSnippets discussing what she learned when creating NFTs from her artwork
- Member discussions:
- IPTC members will be discussing how we might be able to simplify rights management with a cut-down basic set of rights assertions, possibly creating a simpler alternative to RightsML
- IPTC members will also be discussing the News Architecture and how we can better utilise the key data model that underlies both NewsML-G2 and ninjs
- and more!
Attendance to the 2022 IPTC Spring Meeting is free for all delegates and member experts from IPTC member organisations.
Invited speakers are welcome to attend the day on which they are speaking.
IPTC members will be appearing at imaging.org’s Imaging Science and Technology DigiTIPS 2022 meeting series tomorrow, April 26.
The session description is as follows:
Abstract: Learn how embedded photo metadata can aid in a data-driven workflow from capture to publish. Discover what details exist in your images; and learn how you can affix additional information so that you and others can manage your collection of images. See how you can embed info to automatically fill in “Alt Text” to images shown on your website. Explore how you can test your metadata workflow to maximize interoperability.”
Registration is still open. You can register at https://www.imaging.org/Site/IST/Conferences/DigiTIPS/DigiTIPS_Home.aspx?Entry_CCO=3#Entry_CCO
A hot topic in media circles these days is “synthetic media”. That is, media that was created either partly or fully by a computer. Usually the term is used to describe content created either partly or wholly by AI algorithms.
IPTC’s Video Metadata Working Group has been looking at the topic recently and we concluded that it would be useful to have a way to describe exactly what type of content a particular media item is. Is it a raw, unmodified photograph, video or audio recording? Is it a collage of existing photos, or a mix of synthetic and captured content? Was it created using software trained on a set of sample images or videos, or is it purely created by an algorithm?
We have an existing vocabulary that suits some of this need: Digital Source Type. This vocabulary was originally created to be able to describe the way in which an image was scanned into a computer, but it also represented software-created images at a high level. So we set about expanding and modifying that vocabulary to cover more detail and more specific use cases.
It is important to note that we are only describing the way a media object has been created: we are not making any statements about the intent of the user (or the machine) in creating the content. So we deliberately don’t have a term “deepfake”, but we do have “trainedAlgorithmicMedia” which would be the term used to describe a piece of content that was created by an AI algorithm such as a Generative Adversarial Network (GAN).
Here are the terms we propose to include in the new version of the Digital Source Type vocabulary. (New terms and definition changes are marked in bold text. Existing terms are included in the list for clarity.)
|Term name||Original digital capture sampled from real life|
|Term description||The digital media is captured from a real-life source using a digital camera or digital recording device|
|Examples||Digital photo or video taken using a digital SLR or smartphone camera|
|Term name||Digitised from a negative on film|
|Term description||The digital media was digitised from a negative on film on any other transparent medium|
|Examples||Film scanned from a moving image negative|
|Term name||Digitised from a positive on film|
|Term description||The digital media was digitised from a positive on a transparency or any other transparent medium|
|Examples||Digital photo scanned from a photographic positive|
|Term name||Digitised from a print on non-transparent medium|
|Term description||The digital image was digitised from an image printed on a non-transparent medium|
|Examples||Digital photo scanned from a photographic print|
|Term name||Original media with minor human edits|
|Term description||Minor augmentation or correction by a human, such as a digitally-retouched photo used in a magazine|
|Examples||Video camera recording, manipulated digitally by a human editor|
|Term name||Composite of captured elements|
|Term description||Mix or composite of several elements that are all captures of real life|
|Examples||* A composite image created by a digital artist in Photoshop based on several source images
* Edited sequence or composite of video shots
|Term name||Algorithmically-enhanced media|
|Term description||Minor augmentation or correction by algorithm|
|Examples||A photo that has been digitally enhanced using a mechanism such as Google Photos’ “de-noise” feature|
|Term name||Data-driven media|
|Term description||Digital media representation of data via human programming or creativity|
|Examples||A representation of a distant galaxy created by analysing the outputs of a deep-space telescope (as opposed to a regular camera)
An infographic created using a computer drawing tool such as Adobe Illustrator or AutoCAD
|Term name||Digital art|
|Term description||Media created by a human using digital tools|
|Examples||* A cartoon drawn by an artist into a digital tool using a digital pencil, a tablet and a drawing package such as Procreate or Affinity Designer
* A scene from a film/movie created using Computer Graphic Imagery (CGI)
* Electronic music composition using purely synthesised sounds
|Term name||Virtual recording|
|Term description||Live recording of virtual event based on synthetic and optionally captured elements|
|Examples||* A recording of a computer-generated sequence, e.g. from a video game
* A recording of a Zoom meeting
|Term name||Composite including synthetic elements|
|Term description||Mix or composite of several elements, at least one of which is synthetic|
|Examples||* Movie production using a combination of live-action and CGI content, e.g. using Unreal engine to generate backgrounds
* A capture of an augmented reality interaction with computer imagery superimposed on a camera video, e.g. someone playing Pokemon Go
|Term name||Trained algorithmic media|
|Term description||Digital media created algorithmically using a model derived from sampled content|
|Examples||* Image based on deep learning from a series of reference examples
* A “speech-to-speech” generated audio or “deepfake” video using a combination of a real actor and an AI model
* “Text-to-image” using a text input to feed an algorithm that creates a synthetic image
|Term name||Algorithmic media|
|Term description||Media created purely by an algorithm not based on any sampled training data, e.g. an image created by software using a mathematical formula|
|Examples||* A purely computer-generated image such as a pattern of pixels generated mathematically e.g. a Mandelbrot set or fractal diagram
* A purely computer-generated moving image such as a pattern of pixels generated mathematically
We propose that the following term, which exists in the current DigitalSourceType CV, be retired:
|Term ID||RETIRE: softwareImage|
|Term name||Created by software|
|Term description||The digital image was created by computer software|
|Note||We propose that trainedAlgorithmicMedia or algorithmnicMedia be used instead of this term.|
We welcome all feedback from across the industry to these proposed terms.
Anyone who has managed photo metadata can attest that it is often difficult to know which metadata properties to use for different purposes. It is especially tricky to know how to tag consistently across different metadata standards. For example, how should a copyright notice be expressed in Exif, IPTC Photo Metadata and schema.org metadata?
For software vendors wanting to build accurate mapping into their tools to make life easier for their customers, it’s no easier. For a while, a document created by a consortium of vendors known as the Metadata Working Group solved some of the problems, but the MWG Guidelines are no longer available online.
To solve this problem, the IPTC collaborated with Exif experts at CIPA, the camera products industry group that maintains the Exif standard. We also spoke with the team behind schema.org. Based on these conversations, we created a document that describes how to map properties between these formats. The aim is to remove any ambiguity regarding which IPTC Photo Metadata properties are semantically equivalent to Exif tags and schema.org properties.
Generally, Exif tags and IPTC Photo Metadata properties represent different things: Exif mainly represents the technical data around capturing an image, while IPTC focuses on describing the image and its administrative and rights metadata, and schema.org covers expressing metadata in a web page. However, quite a few properties are shared by all standards, such as who is the Creator of the image, the free-text description of what the image shows, or the date when the image was taken. Therefore it is highly recommended to have the same value in the corresponding fields of the different standards.
The IPTC Photo Metadata Mapping Guidelines outlines the 17 IPTC Photo Metadata Standard properties with corresponding fields in Exif and/or Schema.org. Further short textual notes help to implement these mappings correctly.
The intended audience of the document is those managing the use of photo metadata in businesses and the makers of software that handles photo metadata.The IPTC Photo Metadata Mapping Guidelines document can be accessed on the iptc.org website. We encourage IPTC members to provide feedback through the usual channels, and non-members to respond with feedback and questions on the public IPTC Photo Metadata email discussion group.