<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.diglib.org/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Jchapman</id>
	<title>DLF Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.diglib.org/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Jchapman"/>
	<link rel="alternate" type="text/html" href="https://wiki.diglib.org/Special:Contributions/Jchapman"/>
	<updated>2026-04-16T20:21:23Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.44.0</generator>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=8250</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=8250"/>
		<updated>2017-01-25T15:18:48Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a [http://dashboard.diglib.org/ Digitization Cost Calculator] that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current Projects==&lt;br /&gt;
===Digitization Cost Calculator and associated data set===&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 85 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a new version of the calculator complete. [https://dashboard.diglib.org/ The new Digitization Cost Calculator will be available here] on November 7th 2016.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Processes for which the new calculator will provide data&#039;&#039;&#039;&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group solicited enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign. We currently have the minimum necessary; the calculator will be vastly improved if we have more! Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here].&lt;br /&gt;
&lt;br /&gt;
===Publication Information===&lt;br /&gt;
*[https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
*[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit Best practices and guidelines for the collection of time data for digitization processes]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome. If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Membership===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Wayne Graham, CLIR (technical lead on Cost Calculator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Membership===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7836</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7836"/>
		<updated>2016-11-04T05:14:02Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a [https://dashboard.diglib.org/ Digitization Cost Calculator] that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current Projects==&lt;br /&gt;
===Digitization Cost Calculator and associated data set===&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 85 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a new version of the calculator complete. [https://dashboard.diglib.org/ The new Digitization Cost Calculator will be available here] on November 7th 2016.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Processes for which the new calculator will provide data&#039;&#039;&#039;&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group solicited enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign. We currently have the minimum necessary; the calculator will be vastly improved if we have more! Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here].&lt;br /&gt;
&lt;br /&gt;
===Publication Information===&lt;br /&gt;
*[https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
*[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit Best practices and guidelines for the collection of time data for digitization processes]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome. If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Membership===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Wayne Graham, CLIR (technical lead on Cost Calculator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Membership===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7835</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7835"/>
		<updated>2016-11-04T05:13:36Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a Digitization Cost Calculator that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current Projects==&lt;br /&gt;
===Digitization Cost Calculator and associated data set===&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 85 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a new version of the calculator complete. [https://dashboard.diglib.org/ The new Digitization Cost Calculator will be available here] on November 7th 2016.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Processes for which the new calculator will provide data&#039;&#039;&#039;&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group solicited enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign. We currently have the minimum necessary; the calculator will be vastly improved if we have more! Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here].&lt;br /&gt;
&lt;br /&gt;
===Publication Information===&lt;br /&gt;
*[https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
*[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit Best practices and guidelines for the collection of time data for digitization processes]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome. If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Membership===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Wayne Graham, CLIR (technical lead on Cost Calculator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Membership===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7834</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7834"/>
		<updated>2016-11-04T05:11:04Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: updated calculator URL&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. [https://dashboard.diglib.org/ View the Digitization Cost Calculator here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
===Try to purposefully time processes, and not include time for overhead or other activities=== &lt;br /&gt;
People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans the process was performed on -- or, if it is a pre-scanning process, you will need to identify the quantity of material. Later, the “number of scans” is determined by the number of scans that will be produced from this material. For pre-scanning processes, it is easiest to time based on concrete quantities: i.e., an entire collection, a set of boxes.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 20 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch [https://youtu.be/9WUo-0EJxHg this short video] from Northwestern University about how they collect data&lt;br /&gt;
* ... we hope to have more examples soon, let us know if you&#039;d like to contribute something!&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans), and what percentage of materials the process was performed on (for everything other than image capture).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form [https://drive.google.com/open?id=0B4veUYJxzAmSUXFjbU51dm5DUzg here].&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7826</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7826"/>
		<updated>2016-10-20T17:16:53Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a Digitization Cost Calculator that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current Projects==&lt;br /&gt;
===Digitization Cost Calculator and associated data set===&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 80 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a version of the calculator complete. Stay tuned for updates!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Processes for which the new calculator will provide data&#039;&#039;&#039;&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group solicited enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign. We currently have the minimum necessary; the calculator will be vastly improved if we have more! Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here].&lt;br /&gt;
&lt;br /&gt;
===Publication Information===&lt;br /&gt;
*[https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
*[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit Best practices and guidelines for the collection of time data for digitization processes]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome. If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Membership===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Membership===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7819</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7819"/>
		<updated>2016-10-20T16:16:20Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a Digitization Cost Calculator that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current Projects==&lt;br /&gt;
===Digitization Cost Calculator and associated data set===&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 80 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a version of the calculator complete. Stay tuned for updates!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Processes for which the new calculator will provide data&#039;&#039;&#039;&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group solicited enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign. We currently have the minimum necessary; the calculator will be vastly improved if we have more! Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here].&lt;br /&gt;
&lt;br /&gt;
===Publication Information===&lt;br /&gt;
*[https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
*[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit Best practices and guidelines for the collection of time data for digitization processes]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome. If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Membership===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Membershipp===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7818</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7818"/>
		<updated>2016-10-20T16:15:34Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a Digitization Cost Calculator that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current Projects==&lt;br /&gt;
===Digitization Cost Calculator and associated data set===&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 80 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a version of the calculator complete. Stay tuned for updates!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Processes for which the new calculator will provide data&#039;&#039;&#039;&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group solicited enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign. We currently have the minimum necessary; the calculator will be vastly improved if we have more! Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here].&lt;br /&gt;
&lt;br /&gt;
===Publication Information===&lt;br /&gt;
*[https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
*[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit Best practices and guidelines for the collection of time data for digitization processes]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome. If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Members of the Cost Assessment Working Groups===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7814</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7814"/>
		<updated>2016-10-20T15:54:08Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a Digitization Cost Calculator that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current projects==&lt;br /&gt;
===Digitization Cost Calculator and associated data set===&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 80 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a version of the calculator complete. Stay tuned for updates!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Processes for which the new calculator will provide data&#039;&#039;&#039;&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group solicited enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign. We currently have the minimum necessary; the calculator will be vastly improved if we have more! Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here].&lt;br /&gt;
&lt;br /&gt;
==Publication information==&lt;br /&gt;
*[https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
*[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit Best practices and guidelines for the collection of time data for digitization processes]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome. If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Members of the Cost Assessment Working Groups===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7813</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7813"/>
		<updated>2016-10-20T15:52:40Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a Digitization Cost Calculator that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current projects==&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 80 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a version of the calculator complete. Stay tuned for updates!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Processes for which the new calculator will provide data&#039;&#039;&#039;&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group solicited enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign. We currently have the minimum necessary; the calculator will be vastly improved if we have more! Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here].&lt;br /&gt;
&lt;br /&gt;
==Publication information==&lt;br /&gt;
*[https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
*[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit Best practices and guidelines for the collection of time data for digitization processes]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome. If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Members of the Cost Assessment Working Groups===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7812</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7812"/>
		<updated>2016-10-20T15:51:59Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a Digitization Cost Calculator that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current projects==&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 80 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a version of the calculator complete. Stay tuned for updates!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
===Processes for which the new calculator will provide data===&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group solicited enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign. We currently have the minimum necessary; the calculator will be vastly improved if we have more! Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here].&lt;br /&gt;
&lt;br /&gt;
==Publication information==&lt;br /&gt;
*[https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
*[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit Best practices and guidelines for the collection of time data for digitization processes]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome. If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Members of the Cost Assessment Working Groups===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7810</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7810"/>
		<updated>2016-10-20T15:17:55Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
the Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a Digitization Cost Calculator that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current projects==&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 80 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a version of the calculator complete. Stay tuned for updates!&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
===Processes for which the new calculator will provide data (and for which we are collecting data)===&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group is worked to solicit enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign.&lt;br /&gt;
&lt;br /&gt;
Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here]!&lt;br /&gt;
&lt;br /&gt;
If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
See the cost assessment bibliography produced by the Cost Assessment working group here:&lt;br /&gt;
* [https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome, get in touch with your ideas! joyce.chapman@duke.edu.&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Members of the Cost Assessment Working Groups===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7809</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7809"/>
		<updated>2016-10-20T15:15:56Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
the Cost Assessment working group&#039;s tasks are to aggregate and make freely available a large set of data on the time it takes to perform various tasks involved in the digitization process, in order to assist organizations in digitization project planning and benchmarking. To that end, we are building a Digitization Cost Calculator that takes advantage of the contributed data set to provide digitization cost estimates based on various input parameters. Founded in 2014, the Cost Assessment working group of the DLF Assessment Interest Group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are now in use, with the goal of standardizing collection of such data in the field as well as to guide data submissions to the Digitization Cost Calculator. &lt;br /&gt;
&lt;br /&gt;
==Current projects==&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- current status&#039;&#039;: The data definitions authored by the cost assessment working group have now informed a modified structure of the calculator with expanded capabilities that we are completing in 2016 with the assistance of Wayne Graham, Technical Director at CLIR and DLF. We gathered 80 data submissions in summer of 2016 with a Day of Data campaign, spent the fall reviewing and normalizing submissions, and are now close to having a version of the calculator complete. Stay tuned for updates!&lt;br /&gt;
&lt;br /&gt;
==Past projects==&lt;br /&gt;
&#039;&#039;Digitization Cost Calculator -- what&#039;s been completed&#039;&#039;: In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. The original Digitization Cost Calculator was built as a proof of concept and contains data from only a handful of institutions. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was subsequently founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which are used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. &lt;br /&gt;
&lt;br /&gt;
===Processes for which the new calculator will provide data (and for which we are collecting data)===&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form].&lt;br /&gt;
&lt;br /&gt;
In 2016, the group is worked to solicit enough data contributions to build a functional digitization cost calculator 2.0. We needed at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. In order to further this goal, we held a Day of Data campaign.&lt;br /&gt;
&lt;br /&gt;
Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here]!&lt;br /&gt;
&lt;br /&gt;
If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
See the cost assessment bibliography produced by the Cost Assessment working group here:&lt;br /&gt;
* [https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
&lt;br /&gt;
==Get Involved==&lt;br /&gt;
The Cost Assessment working group welcomes new members. Ongoing member duties include quarterly review of recent Cost Calculator data submissions. Those who would like to work on new projects or expansions of existing projects are welcome, get in touch with your ideas! joyce.chapman@duke.edu.&lt;br /&gt;
&lt;br /&gt;
===2015-2016 Members of the Cost Assessment Working Groups===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California, Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7731</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7731"/>
		<updated>2016-06-01T15:39:40Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* How are other people doing it?! */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
===Try to purposefully time processes, and not include time for overhead or other activities=== &lt;br /&gt;
People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans the process was performed on -- or, if it is a pre-scanning process, you will need to identify the quantity of material. Later, the “number of scans” is determined by the number of scans that will be produced from this material. For pre-scanning processes, it is easiest to time based on concrete quantities: i.e., an entire collection, a set of boxes.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 20 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch [https://youtu.be/9WUo-0EJxHg this short video] from Northwestern University about how they collect data&lt;br /&gt;
* ... we hope to have more examples soon, let us know if you&#039;d like to contribute something!&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans), and what percentage of materials the process was performed on (for everything other than image capture).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form [https://drive.google.com/open?id=0B4veUYJxzAmSUXFjbU51dm5DUzg here].&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7725</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7725"/>
		<updated>2016-05-24T20:46:48Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* How to collect data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
===Try to purposefully time processes, and not include time for overhead or other activities=== &lt;br /&gt;
People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans the process was performed on -- or, if it is a pre-scanning process, you will need to identify the quantity of material. Later, the “number of scans” is determined by the number of scans that will be produced from this material. For pre-scanning processes, it is easiest to time based on concrete quantities: i.e., an entire collection, a set of boxes.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 20 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* ... we hope to have more examples soon, let us know if you&#039;d like to contribute something!&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans), and what percentage of materials the process was performed on (for everything other than image capture).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form [https://drive.google.com/open?id=0B4veUYJxzAmSUXFjbU51dm5DUzg here].&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7724</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7724"/>
		<updated>2016-05-24T20:46:15Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* How to collect data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
===Try to purposefully time processes, and not include time for overhead or other activities.=== P&lt;br /&gt;
eople using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans the process was performed on -- or, if it is a pre-scanning process, you will need to identify the quantity of material. Later, the “number of scans” is determined by the number of scans that will be produced from this material. For pre-scanning processes, it is easiest to time based on concrete quantities: i.e., an entire collection, a set of boxes.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 20 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* ... we hope to have more examples soon, let us know if you&#039;d like to contribute something!&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans), and what percentage of materials the process was performed on (for everything other than image capture).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form [https://drive.google.com/open?id=0B4veUYJxzAmSUXFjbU51dm5DUzg here].&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7720</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7720"/>
		<updated>2016-05-24T14:26:57Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* How to submit data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Try to purposefully time processes, and not include time for overhead or other activities.&#039;&#039;&#039; People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans the process was performed on -- or, if it is a pre-scanning process, you will need to identify the quantity of material. Later, the “number of scans” is determined by the number of scans that will be produced from this material. For pre-scanning processes, it is easiest to time based on concrete quantities: i.e., an entire collection, a set of boxes.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 20 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* ... we hope to have more examples soon, let us know if you&#039;d like to contribute something!&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans), and what percentage of materials the process was performed on (for everything other than image capture).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form here.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7719</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7719"/>
		<updated>2016-05-24T14:26:14Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* How are other people doing it?! */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Try to purposefully time processes, and not include time for overhead or other activities.&#039;&#039;&#039; People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans the process was performed on -- or, if it is a pre-scanning process, you will need to identify the quantity of material. Later, the “number of scans” is determined by the number of scans that will be produced from this material. For pre-scanning processes, it is easiest to time based on concrete quantities: i.e., an entire collection, a set of boxes.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 20 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* ... we hope to have more examples soon, let us know if you&#039;d like to contribute something!&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form here.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7718</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7718"/>
		<updated>2016-05-24T14:24:25Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* Methods of timing */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Try to purposefully time processes, and not include time for overhead or other activities.&#039;&#039;&#039; People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans the process was performed on -- or, if it is a pre-scanning process, you will need to identify the quantity of material. Later, the “number of scans” is determined by the number of scans that will be produced from this material. For pre-scanning processes, it is easiest to time based on concrete quantities: i.e., an entire collection, a set of boxes.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 20 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* See [https://docs.google.com/document/d/1hM607wGfnSm8gSQa9RZWzT_tmv-g89j9x3Txt_1fscg/edit this data collection form] contributed by McGill University&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form here.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7717</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7717"/>
		<updated>2016-05-24T14:23:11Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* Methods of timing */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Try to purposefully time processes, and not include time for overhead or other activities.&#039;&#039;&#039; People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans the process was performed on -- or, if it is a pre-scanning process, you will need to identify the quantity of material. Later, the “number of scans” is determined by the number of scans that will be produced from this material. For pre-scanning processes, it is easiest to time based on concrete quantities: i.e., an entire collection, a set of boxes.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20-30 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is very difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 30 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* See [https://docs.google.com/document/d/1hM607wGfnSm8gSQa9RZWzT_tmv-g89j9x3Txt_1fscg/edit this data collection form] contributed by McGill University&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form here.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7716</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7716"/>
		<updated>2016-05-24T14:19:10Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* How to collect data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Try to purposefully time processes, and not include time for overhead or other activities.&#039;&#039;&#039; People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans. In the case of a process performed before image capture, the “number of scans” is determined by the number of scans that will be produced -- if 100 manuscript pages are reviewed that will be scanned on both sides, call it about 200 scans. You also need to know this so you can estimate the percentage of materials the process took place on -- 10 fasteners removed for 200 scans is 5%.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20-30 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is very difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 30 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* See [https://docs.google.com/document/d/1hM607wGfnSm8gSQa9RZWzT_tmv-g89j9x3Txt_1fscg/edit this data collection form] contributed by McGill University&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form here.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7715</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7715"/>
		<updated>2016-05-24T14:14:08Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* How to collect data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for multiple processes&lt;br /&gt;
* Is just a broad/loose estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Try to purposefully time processes, and not include time for overhead or other activities.&#039;&#039;&#039; People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans. In the case of a process performed before image capture, the “number of scans” is determined by the number of scans that will be produced -- if 100 manuscript pages are reviewed that will be scanned on both sides, call it about 200 scans. You also need to know this so you can estimate the percentage of materials the process took place on -- 10 fasteners removed for 200 scans is 5%.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20-30 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is very difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 30 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* See [https://docs.google.com/document/d/1hM607wGfnSm8gSQa9RZWzT_tmv-g89j9x3Txt_1fscg/edit this data collection form] contributed by McGill University&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form here.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7714</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7714"/>
		<updated>2016-05-24T14:13:08Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* How to collect data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields in which you want to submit data. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for more than one process&lt;br /&gt;
* Is a broad estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Try to purposefully time processes, and not include time for overhead or other activities.&#039;&#039;&#039; People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans. In the case of a process performed before image capture, the “number of scans” is determined by the number of scans that will be produced -- if 100 manuscript pages are reviewed that will be scanned on both sides, call it about 200 scans. You also need to know this so you can estimate the percentage of materials the process took place on -- 10 fasteners removed for 200 scans is 5%.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20-30 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is very difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 30 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* See [https://docs.google.com/document/d/1hM607wGfnSm8gSQa9RZWzT_tmv-g89j9x3Txt_1fscg/edit this data collection form] contributed by McGill University&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form here.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7713</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7713"/>
		<updated>2016-05-24T14:12:26Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: /* How are other people doing it?! */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields you want to submit data for. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for more than one process&lt;br /&gt;
* Is a broad estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Try to purposefully time processes, and not include time for overhead or other activities.&#039;&#039;&#039; People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans. In the case of a process performed before image capture, the “number of scans” is determined by the number of scans that will be produced -- if 100 manuscript pages are reviewed that will be scanned on both sides, call it about 200 scans. You also need to know this so you can estimate the percentage of materials the process took place on -- 10 fasteners removed for 200 scans is 5%.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20-30 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is very difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 30 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video (coming soon) from Northwestern University about how they collect data&lt;br /&gt;
* See [https://docs.google.com/document/d/1hM607wGfnSm8gSQa9RZWzT_tmv-g89j9x3Txt_1fscg/edit this data collection form] contributed by McGill University&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form here.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7712</id>
		<title>Data Collection and Contribution, Digitization Cost Calculator</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Data_Collection_and_Contribution,_Digitization_Cost_Calculator&amp;diff=7712"/>
		<updated>2016-05-23T21:09:58Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: Created page with &amp;quot;&amp;#039;&amp;#039;&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page&amp;#039;&amp;#039;  Digitization is a costly business -- estimating expenses associated...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Assessment:Costs|&amp;lt;&amp;lt;Return to the DLF Assessment Interest Group, Cost Assessment Working Group page]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Digitization is a costly business -- estimating expenses associated with a given digitization project, a fiscal year, or for a grant application, can feel disconnected from the reality of staffing, timelines, and true project costs. The Digital Library Federation’s Assessment Interest Group is developing a Digitization Cost Calculator that runs on data contributed by participating institutions. The calculator will help with digitization project planning by using contributed data to produce average estimates of costs and time for various aspects of the digitization process. You can view the wireframes for the calculator [https://drive.google.com/folderview?id=0B4veUYJxzAmSbUIzODJDZHc3MEE&amp;amp;usp=drive_web here], or use a beta version (simpler than the final version) [http://statelibrarync.org/plstats/digitization_calculator.php here]. &lt;br /&gt;
&lt;br /&gt;
But we need your help! We can’t build the calculator unless the community contributes data. You can contribute data at any time for one or more of the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit data fields in the calculator]. &lt;br /&gt;
&lt;br /&gt;
==How to collect data==&lt;br /&gt;
Before you begin, read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes] carefully and make sure that your process conforms to the guidelines for the fields you want to submit data for. Remember that we cannot accept data that:&lt;br /&gt;
* Does not conform with the DLF-AIG [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit definitions of processes]&lt;br /&gt;
* Does not provide both time data and the number of scans to which time data relates&lt;br /&gt;
* Combines time data that cannot be separated for more than one process&lt;br /&gt;
* Is a broad estimate of time data rather than true data collection&lt;br /&gt;
&lt;br /&gt;
Understand that all data is submitted in increments of “per 100 scans.” This is necessary for the calculator to make data usable between institutions. Even if you are submitting data for a process that is not performed “by scan” (i.e., fastener removal, disbinding) you’ll need to pay attention to how many digital images are captured from the quantity of material that you timed staff on as they performed the action, so that it can be calculated as “x time per 100 scans.”&lt;br /&gt;
&lt;br /&gt;
Additionally, for everything other than image capture, you’ll need to submit data on the percentage of materials on which a process was performed. See example 2 below.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 1:&#039;&#039;&#039; A student is assigned to scan glass plate negatives on a flatbed scanner. After one hour, the student reported that they had scanned 20 plates. &lt;br /&gt;
&lt;br /&gt;
Divide 60 (minutes) by 20 (scans) = 3 minutes per scan. &lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
3 (minutes per scan) * 100 (scans) = 300 minutes to complete 100 scans at the rate determined by your session.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Example 2:&#039;&#039;&#039; A staff person is removing staples. In 2 hours of removal, the staff person removes 40 fasteners from papers in a single box of materials. Once the box is scanned, the materials produce 150 scans.&lt;br /&gt;
&lt;br /&gt;
Turn 2 hours into minutes: 2*60 = 120&lt;br /&gt;
Divide 120 (minutes) by 150 (scans) = 0.8 minutes per scan&lt;br /&gt;
&lt;br /&gt;
In this example, to determine per 100 scans: &lt;br /&gt;
&lt;br /&gt;
0.8 (minutes per scan) * 100 (scans) = 80 minutes to complete fastener removal for 100 scans at the rate determined by your session. &lt;br /&gt;
&lt;br /&gt;
Additionally, you’ll report that the process was performed on 27% of materials (40 staples divided by 150 scans evens out to staples in 27% of materials, when “scans” are used as the base material).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Try to purposefully time processes, and not include time for overhead or other activities.&#039;&#039;&#039; People using the Cost Calculator will build in their own overhead time and costs -- it shouldn’t be included in the data submissions. We would rather have you accurately time a student’s scanning pace for a single 3-hour shift than use a larger quantity of unspecific data to make an inaccurate estimate. For example, if you know that a student scanned 1,000 pages during 50 hours of shifts and they were mostly scanning when they worked, you might be inclined to submit scanning data as “20 scans per hour.” In reality this student may have spent 25% of their time doing other tasks, being out sick, answering emails, eating lunch, or going to the restroom. Submitting such inaccurate data from rough estimates can have a huge negative impact on the ability of the Digitization Cost Calculator to help others estimate time and money for projects.&lt;br /&gt;
&lt;br /&gt;
===Methods of timing===&lt;br /&gt;
* Have staff write down a beginning and end time for a shift or stretch of time that includes only one process (e.g., intellectual property review). They will also write down the number of times the process was performed and the number of scans. In the case of a process performed before image capture, the “number of scans” is determined by the number of scans that will be produced -- if 100 manuscript pages are reviewed that will be scanned on both sides, call it about 200 scans. You also need to know this so you can estimate the percentage of materials the process took place on -- 10 fasteners removed for 200 scans is 5%.&lt;br /&gt;
* Give staff stopwatches (or use stopwatch apps on smartphones) to collect data for short periods of defined work. This is particularly useful for processes that are performed intermittently, e.g. disbinding, supporting, flattening, quality control, alignment/rotation -- these may be performed in between bouts of image capture. In such cases, the ability to easily stop and start a stopwatch instead of keeping track of many start and stop times is very useful. You’ll just need to also make sure you’re marking down the number of times you conduct the process and how many total scans were produced per that number of occurrences. &lt;br /&gt;
* Time a process 20-30 times, then create an average. From then on, just count the number of times you perform the process and at the end, multiply the number by the average time you previously calculated. This method lends itself well to a process that is very difficult to time, or that once timed, typically takes the same amount of time again. For example, if you typically perform image cropping after each scan instead of reviewing and cropping a large number of scans at a single time, you might time yourself with a stopwatch 30 times, see that the times are all pretty similar, take the average, and then just mark how many times you perform the process. If it takes an average of 10 seconds to crop an image and you performed it 500 times over a month on a collection of 700 scans, you’ll get your time data by multiplying 500*10, converting it to minutes and then dividing it by 7 to get minutes per 100 scans.&lt;br /&gt;
&lt;br /&gt;
===How are other people doing it?!===&lt;br /&gt;
Some folks who have already contributed data have provided some helpful tips. &lt;br /&gt;
* Watch this short video from Northwestern University about how they collect data&lt;br /&gt;
* See [https://docs.google.com/document/d/1hM607wGfnSm8gSQa9RZWzT_tmv-g89j9x3Txt_1fscg/edit this data collection form] contributed by McGill University&lt;br /&gt;
&lt;br /&gt;
==How to submit data==&lt;br /&gt;
Data is submitted via an [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 online form], which you can [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] before beginning the data submission process. If you have questions, you can contact joyce.chapman@duke.edu. In order to submit data, you’ll need to know which of the calculator fields you are submitting data for (e.g., fastener removal, image stitching, image capture), your time data, and the number of scans the time data pertains to (e.g., 200 minutes to perform process X for 100 scans).&lt;br /&gt;
&lt;br /&gt;
Get your data ready for the submission form: We highly recommend reviewing the [https://drive.google.com/file/d/0B4veUYJxzAmSUXFjbU51dm5DUzg/view preview] of the data submission form before beginning data submission. &lt;br /&gt;
&lt;br /&gt;
For descriptive metadata creation and quality control you will have to determine which sub-category your data falls into: level 1, 2, or 3. Read the [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit field definitions] for guidelines. &lt;br /&gt;
&lt;br /&gt;
For image capture data, you’ll need to be able to specify the image capture device. &lt;br /&gt;
&lt;br /&gt;
For every data field except image capture, you’ll need to estimate what percentage of the overall material a process was performed on; e.g., if you are submitting data in the image “clean up/dust removal” category and it took you an hour to do dust removal required on 20 of the 100 scans, you’ll need to submit the number of minutes that process took (60), the total number of scans (100), and the percentage of scans the process was performed on (20%, or 20/100). Similarly if you performed fastener removal and you estimate there was a staple every 5-pages or so, you’d provide the percentage as either 20% if the  material is one sided (1 staple for an average of every 5 pages, 5 scans) or 10% if the material is two-sided (1 staple for an average of every 5 pages, 10 scans).&lt;br /&gt;
&lt;br /&gt;
For every data submission, you’ll also be asked to submit:&lt;br /&gt;
* Your name and contact information&lt;br /&gt;
* The dates during which data was collected (“March 2016,” “Apr 15-Sept 20 2015,” etc.)&lt;br /&gt;
* The type of material (manuscript, photograph, etc.)&lt;br /&gt;
* The time period of the materials (19th century, 20th century, etc.)&lt;br /&gt;
&lt;br /&gt;
==FAQ==&lt;br /&gt;
&#039;&#039;&#039;How do I contribute data?&#039;&#039;&#039; Use [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 this online form] or contact joyce.chapman@duke.edu. See a Word document preview of the online form here.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What will I need to do?&#039;&#039;&#039; You will time yourself, staff, or student as you/they perform digitization processes during a period of time of your choosing -- a week, a month, the duration of a specific project, etc. The areas in which you can contribute time data include image capture, descriptive metadata creation, quality control, various preparation processes such as condition review, rebinding, formatting, and various post-processing processes such as alignment/rotation, image cropping, and stitching.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Do I have to give you salary/benefits data?&#039;&#039;&#039; No! You only submit time data (how many minutes it took to perform a specific task per 100 scans). “Costs” are estimated using salary/benefits data that people enter when using the calculator. The calculator does not track or store cost data anywhere.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we don’t perform all of the tasks mentioned?&#039;&#039;&#039; That’s fine because the calculations are broken down by task – you only submit data for the specific tasks that you choose. Contributing whatever pieces of your process that are trackable in whatever increments you can track them in, is still incredibly helpful! &lt;br /&gt;
When will I need to do it? Submissions are ongoing.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) time data for process X and process Y smushed together in one number?&#039;&#039;&#039; Unfortunately, we cannot use data that combines multiple processes -- your time data contributions will have to be separate for each process you contribute data for. Alternatively, you can collect sample data for one of the processes and then estimate the aggregate data out into pieces.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;How will my institution benefit?&#039;&#039;&#039; You will have contributed to the creation of a freely-available tool (the Digitization Cost Calculator) that allows users to input their institution’s salary and benefits data, the amount of material being digitized, select which processes they will be undertaking, and then outputs cost and time data based on all aggregate contributed data. This tool will help many organizations in planning future projects and in articulating the true costs of digitization projects.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Will the information I contribute be associated with my institution?&#039;&#039;&#039; Sort of. The data you submit will be aggregated by the calculator with all other data submissions and displayed as part of an average on the results screen when people use the calculator: No individual institution’s information will be discernible in the calculator. However, individual institutional data will be shown on the [http://statelibrarync.org/plstats/data-notes.php Notes About Data] webpage, another part of the Digitization Cost Calculator website. This allows calculator users to get a feel for the wide variation in time and in practice from institution to institution and project to project. Seeing the data apart from the aggregate average can also be helpful if a user feels their institution is more similar to one or more other institutions in the list, and allows them to calculate custom time estimates. The time period over which the data contribution was collected will also be displayed on the Notes About Data page.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if I have some historical digitization data to contribute now?&#039;&#039;&#039; Great, we’d love to have your historical data! Submit your data using the [https://duke.qualtrics.com/jfe/form/SV_3OtqSEAbpl2QDl3 Data Submission Form], or send an email to joyce.chapman@duke.edu, subject line Cost Calculator, to learn more.&lt;br /&gt;
&#039;&#039;&#039;&lt;br /&gt;
What if my historical data is in a different format than the cost calculator data submission requirement?&#039;&#039;&#039; Email joyce.chapman@duke.edu, and we will help migrate the data into the right format. You will have to know how many scans are associated with a block of time, so if there is no way to go back and get that information, we cannot accept time data without knowing what quantity of material it is for.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;What if we track(ed) only a part of the data you are looking for in our digitization workflow?&#039;&#039;&#039; That’s fine, and still very valuable. You can contribute just one piece of data -- you don’t have to have all the fields represented in the calculator.&lt;br /&gt;
I have more questions! Please feel free to contact Joyce Chapman with any additional questions about the project, being a contributor, or using the calculator: joyce.chapman@duke.edu, 919-660-5889.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7711</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7711"/>
		<updated>2016-05-23T20:29:05Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;[[Data Collection and Contribution, Digitization Cost Calculator|Click here for instructions on collecting/contributing data to the Digitization Cost Calculator]]&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which will be used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. See the bottom of the page for a full list of processes for which the calculator will be able to provide data.  &lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form]. The Digitization Cost Calculator 2.0 cannot be built until at least one set of data has been contributed in each of the 20 categories listed in the guidelines.&lt;br /&gt;
&lt;br /&gt;
In 2016, the group is working to solicit enough data contributions to build a functional digitization cost calculator 2.0. This means we need at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. &lt;br /&gt;
&lt;br /&gt;
Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here]!&lt;br /&gt;
&lt;br /&gt;
If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
See the cost assessment bibliography produced by the Cost Assessment working group here:&lt;br /&gt;
* [https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2016-2017 Members of the Cost Assessment Working Groups===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
==Processes for which the calculator will provide data (and for which we are collecting data!)==&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7710</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7710"/>
		<updated>2016-05-23T20:20:09Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which will be used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. See the bottom of the page for a full list of processes for which the calculator will be able to provide data.  &lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September 2015, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form]. The Digitization Cost Calculator 2.0 cannot be built until at least one set of data has been contributed in each of the 20 categories listed in the guidelines.&lt;br /&gt;
&lt;br /&gt;
In 2016, the group is working to solicit enough data contributions to build a functional digitization cost calculator 2.0. This means we need at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. &lt;br /&gt;
&lt;br /&gt;
Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here]!&lt;br /&gt;
&lt;br /&gt;
If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
See the cost assessment bibliography produced by the Cost Assessment working group here:&lt;br /&gt;
* [https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2016-2017 Members of the Cost Assessment Working Groups===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
==Processes for which the calculator will provide data (and for which we are collecting data!)==&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7709</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=7709"/>
		<updated>2016-05-23T20:19:44Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;In 2014 the Cost Assessment working group developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which will be used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project. See the bottom of the page for a full list of processes for which the calculator will be able to provide data.  &lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form]. The Digitization Cost Calculator 2.0 cannot be built until at least one set of data has been contributed in each of the 20 categories listed in the guidelines.&lt;br /&gt;
&lt;br /&gt;
In 2016 year, the group is working to solicit enough data contributions to build a functional digitization cost calculator 2.0. This means we need at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. &lt;br /&gt;
&lt;br /&gt;
Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here]!&lt;br /&gt;
&lt;br /&gt;
If you are interested in helping us collect data for the Digitization Cost Calculator, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
See the cost assessment bibliography produced by the Cost Assessment working group here:&lt;br /&gt;
* [https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
&lt;br /&gt;
===2014-2015 Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Rissmeyer, University of California Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
===2016-2017 Members of the Cost Assessment Working Groups===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Jane Davis, Saint Louis Federal Reserve Bank&lt;br /&gt;
* Ann Hanlon, University of Wisconsin Milwaukee&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Genya O&#039;Gara, VIVA&lt;br /&gt;
* Chrissy Rissmeyer, University of California Santa Barbara&lt;br /&gt;
* Sarah Severson, McGill University Library&lt;br /&gt;
* Emily Shaw, Ohio State University Libraries&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
==Processes for which the calculator will provide data (and for which we are collecting data!)==&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:User_Studies&amp;diff=259</id>
		<title>Assessment:User Studies</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:User_Studies&amp;diff=259"/>
		<updated>2015-10-19T15:07:41Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The User Studies Working Group compiles resources, conducts research, and drafts literature to assist those who are interesting in evaluating users of digital repositories and their needs.&lt;br /&gt;
&lt;br /&gt;
== The User Studies Working Group white paper ==&lt;br /&gt;
&lt;br /&gt;
Participants identified four assessment topics to explore during the [https://docs.google.com/document/d/1rLdMSPCbR_F-hE0ZhWzS70IBGEMh1450ff0x_ioPrIk/edit Assessment Breakout Session] at the 2014 DLF Forum:&lt;br /&gt;
&lt;br /&gt;
# User/Usability studies&lt;br /&gt;
# Identifying users and uses&lt;br /&gt;
# Return On Investment (ROI)&lt;br /&gt;
# Reuse of content/what do users do with content&lt;br /&gt;
&lt;br /&gt;
These four groups will make up the focus of a white paper on assessing users and user needs.  The working group will complete a series of tasks to generate the white paper:&lt;br /&gt;
&lt;br /&gt;
# Compile a [https://docs.google.com/document/d/1AVlT3QlmhbcUmoRtZ0hpvEhpbD9_Eld5bHsa-JNHX24/edit?usp=sharing bibliography]&lt;br /&gt;
# Develop a literature review -- &#039;&#039;&#039;Completed&#039;&#039;&#039;&lt;br /&gt;
# Release [https://docs.google.com/document/d/1Jd7N6BMh7g1DbgSpL0vigY4s87k7UQjOJLvgYYkqP0c/edit?usp=sharing &#039;&#039;&#039;Surveying the Landscape: Use and Usability Assessment of Digital Libraries&#039;&#039;&#039;] for public comment&lt;br /&gt;
&lt;br /&gt;
The final draft of the white paper will be released in December 2015. &lt;br /&gt;
&lt;br /&gt;
=== White Paper Participants ===&lt;br /&gt;
&lt;br /&gt;
==== White Paper Team ====&lt;br /&gt;
* Joyce Chapman, Duke University&lt;br /&gt;
* Megan Hurst, Athenaeum21 Consulting&lt;br /&gt;
* Elizabeth Joan Kelly, Loyola University New Orleans&lt;br /&gt;
* Caroline Muglia, University of Southern California&lt;br /&gt;
* Genya O’Gara, Virtual Library of Virginia&lt;br /&gt;
* Ayla Stein, University of Illinois at Urbana-Champaign&lt;br /&gt;
* Santi Thompson, University of Houston (Co-Leader)&lt;br /&gt;
* Liz Woolcott, Utah State University (Co-Leader)&lt;br /&gt;
* Tao Zhang, Purdue University&lt;br /&gt;
&lt;br /&gt;
==== Literature Review Team ====&lt;br /&gt;
* Joyce Chapman, Duke University&lt;br /&gt;
* Jody DeRidder, University of Alabama&lt;br /&gt;
* Elizabeth Joan Kelly, Loyola University New Orleans&lt;br /&gt;
* Martha Kyrillidou, Association of Research Libraries&lt;br /&gt;
* Caroline Muglia, University of Southern California&lt;br /&gt;
* Genya O’Gara, Virtual Library of Virginia&lt;br /&gt;
* Ayla Stein, University of Illinois at Urbana-Champaign&lt;br /&gt;
* Santi Thompson, University of Houston (Leader)&lt;br /&gt;
* Rachel Trent, State Library of North Carolina&lt;br /&gt;
* Liz Woolcott, Utah State University&lt;br /&gt;
&lt;br /&gt;
==== Bibliography Team ====&lt;br /&gt;
* Elizabeth Joan Kelly, Loyola University New Orleans&lt;br /&gt;
* Santi Thompson, University of Houston (Leader)&lt;br /&gt;
* Rachel Trent, George Washington University&lt;br /&gt;
* Liz Woolcott, Utah State University&lt;br /&gt;
* Tao Zhang, Purdue University&lt;br /&gt;
&lt;br /&gt;
== Possible Next Steps ==&lt;br /&gt;
&lt;br /&gt;
Future efforts from the group may include generating complementary content, such as:&lt;br /&gt;
# Templates/guides &lt;br /&gt;
# Toolkits&lt;br /&gt;
# Original Research&lt;br /&gt;
&lt;br /&gt;
== Get Involved ==&lt;br /&gt;
&lt;br /&gt;
The working group maintains a [https://drive.google.com/folderview?id=0BxKlOQm6nnVndnJrR200UFZoSm8&amp;amp;usp=sharing Google Drive folder], which contains meeting notes and drafts of working documents. &lt;br /&gt;
&lt;br /&gt;
If you have any questions, or are interested in participating in the working group, please contact [mailto:sathompson3@uh.edu Santi Thompson]. You can also fill out our [https://docs.google.com/spreadsheets/d/16X0B-NSulh1V6HuqkSZZq0El35AQlZLbpWUlzNx2Puo/edit User Studies Participant] spreadsheet.&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Citations&amp;diff=258</id>
		<title>Assessment:Citations</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Citations&amp;diff=258"/>
		<updated>2015-10-19T15:07:22Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;__TOC__&lt;br /&gt;
&lt;br /&gt;
== Introduction ==&lt;br /&gt;
&lt;br /&gt;
The DLF Assessment Interest Group (DLF AIG) citations working group formed in the Fall of 2014 following the successful launch of the larger DLF AIG at the 2014 DLF forum. Volunteers interested in developing best practices for creating and displaying citations for digital library assets were solicited through the Digital Library Assessment Google Group. The citations working group was tasked with answering the following questions: &lt;br /&gt;
&lt;br /&gt;
# What should a citation consist of?&lt;br /&gt;
# How can we best support appropriate citations?&lt;br /&gt;
# To what extent do common citation formats support this? &lt;br /&gt;
# What are the limitations of current digital library software systems for displaying citation information? &lt;br /&gt;
# What are best practices for displaying citation information for reference manager software capture? &lt;br /&gt;
&lt;br /&gt;
The citations working group was further split up to accomplish the following goals:&lt;br /&gt;
# Draft citation standards, based on what can and can&#039;t be incorporated into APA, Chicago, and MLA, that incorporate the necessary elements for digitized special collections and institutional repository content&lt;br /&gt;
# Explore how best to display citation information in page markup so reference managers like Mendeley, Zotero, and EndNote Web can easily parse the pages to import complete citation information&lt;br /&gt;
# Compile information on the technical issues for displaying citations from major software systems&lt;br /&gt;
&lt;br /&gt;
As of DLF 2015, the first goal involving draft citation standards has been completed. An initial draft was circulated to the DLF Assessment group using Google Docs from March 6 through April 3, 2015. At that point comments and edits that were grammar and/or punctuation-related as well as suggestions to clarify text were resolved, and a new document was moved to for further editing. Additional comments were invited through April 24, 2015. The resulting white paper, &amp;quot;Guidelines for Citing Library-hosted, Unique Digital Assets,&amp;quot; is now available.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Access to the [https://docs.google.com/document/d/1eAb9ObJ_KU6Lp9_3vngOehS16XVX4vaW9B6_gNQ52_M/edit?usp=sharing white paper is available via Google docs here,]  and will be posted on Figshare following DLF 2015.&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
== Abstract of &amp;quot;Guidelines for Citing Library-hosted, Unique Digital Assets&amp;quot; ==&lt;br /&gt;
&lt;br /&gt;
These draft guidelines were developed in response to a call for steps forward made by the Digital Library Federation (DLF) Assessment subgroup on Benefits. One of the areas identified for development included the establishment of best practices and guidelines for citing digital libraries. Most if not all major citation styles do not provide direction for citing digital objects, and it is difficult for the administrators of digital repositories to use traditional citation metrics to track the use of their digital objects in scholarly output. Scholarly literature has recently begun to focus on challenges and best practices for citing data sets. In addition, some citation styles, as well as individual institutions, provide guidelines for citing special collections materials. Much of this work can be drawn upon in order to format citations for digitized special collection and cultural heritage materials (e.g. rare books, manuscript materials, images, moving images, etc.) and institutional repository content. These guidelines attempt to address the following questions:&lt;br /&gt;
&lt;br /&gt;
# What should a citation consist of?&lt;br /&gt;
# How can we best support appropriate citations?&lt;br /&gt;
# To what extent do common citation formats support this?&lt;br /&gt;
&lt;br /&gt;
This document outlines suggested citation guidelines in an attempt to fill in gaps when the citation style does not give guidelines for a type of source—in this case, institutional repository and digitized cultural heritage objects. &lt;br /&gt;
&lt;br /&gt;
== White Paper Author and Contributors ==&lt;br /&gt;
&lt;br /&gt;
* Elizabeth Joan Kelly, Loyola University New Orleans (author)&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Acknowledgements&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
The following people provided valuable feedback, edits, and comments on the various drafts of this document:&lt;br /&gt;
* Geoffrey Bilder&lt;br /&gt;
* Bianca Crowley&lt;br /&gt;
* Jody DeRidder&lt;br /&gt;
* Kevin Hawkins&lt;br /&gt;
* Stacy Konkiel&lt;br /&gt;
* Martha Kyrillidou&lt;br /&gt;
* Bill Landis&lt;br /&gt;
* Elliot D. Williams&lt;br /&gt;
&lt;br /&gt;
== Next Steps ==&lt;br /&gt;
&lt;br /&gt;
This document will be presented and discussed at the 2015 Digital Library Federation Forum as part of the session  &amp;quot;Collaborative Efforts to Develop Best Practices in Assessment: A Progress Report&amp;quot; on Monday, October 26 at 1:30pm Pacific Time. The session will be [http://www.diglib.org/forums/2015forum/livestream-schedule available via livestreaming] during the conference. There will also be an opportunity to learn more about and discuss the group&#039;s work at the DLF Assessment lunch on Tuesday October 27.&lt;br /&gt;
&lt;br /&gt;
Other groups both within the Digital Library Federation and without are working on how best to display citation information in page markup so reference managers, local library catalogs, discovery systems, and others can easily parse digital repositories to import complete citation information. Those interested in participating in developing best practices for creating and displaying citations of library-hosted, unique digital assets should contribute to the discussion through the [https://groups.google.com/forum/#!forum/digital-library-assessment DLF AIG Google Group] or in person at DLF.&lt;br /&gt;
&lt;br /&gt;
Please direct questions about the [https://docs.google.com/document/d/1eAb9ObJ_KU6Lp9_3vngOehS16XVX4vaW9B6_gNQ52_M/edit?usp=sharing &amp;quot;Guidelines for citing library-hosted, unique digital assets&amp;quot;] to Elizabeth Joan Kelly at [mailto:ejkelly@loyno.edu ejkelly at loyno.edu].&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Analytics&amp;diff=257</id>
		<title>Assessment:Analytics</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Analytics&amp;diff=257"/>
		<updated>2015-10-19T15:07:02Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The DLF Assessment Interest Group (DLF AIG) analytics working group first formed in the Fall of 2014 following the successful launch of the larger DLF AIG and the DLF fall forum.  In December 2014, the DLF AIG analytics working group set out to draft recommendations for using web analytics for assessment in digital libraries.   The group eventually decided to scope their efforts around a subset of the metrics in the widely adopted Google Analytics service.  Library-centric examples were also gathered to support each recommended metric. The resulting white paper, &amp;quot;Best Practices for Google Analytics in Digital Libraries&amp;quot; is now available.   The white paper is intended for digital library managers, and curators who want to use analytics to understand more about users of, access to, and use of digital library materials. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Access to the [https://docs.google.com/document/d/1QmiLJEZXGAY-s7BG_nyF6EUAqcyH0mhQ7j2VPpLpxCQ/edit?usp=sharing white paper is available via Google docs here], and will be posted on Figshare following DLF 2015.&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
== Abstract of: &amp;quot;Best Practices for Google Analytics in Digital Libraries&amp;quot; ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The purpose of this white paper is to provide digital libraries with guidelines that maximize the effectiveness and relevance of data collected through the Google Analytics service for assessment purposes. The document recommends tracking 14 specific metrics within Google Analytics, and provides library-centric examples of how to employ the resulting data in making decisions and setting institutional goals and priorities. The guidelines open with a literature review, and also include theoretical and structural methods for approaching analytics data gathering, examples of platform specific implementation considerations, Google Analytics set-up tips and terminology, as well as recommended resources for learning more about web analytics. The DLF Assessment Interest Group Analytics working group, which produced this white paper, looks forward to receiving feedback and additional examples of using the recommended metrics for digital library assessment activities. &lt;br /&gt;
&lt;br /&gt;
== White Paper Authors ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Molly Bragg (Co-coordinator of the Analytics Working Group), Duke University Libraries&lt;br /&gt;
* Joyce Chapman (Co-coordinator of the Analytics Working Group), Duke University Libraries&lt;br /&gt;
* Jody DeRidder, University of Alabama Libraries&lt;br /&gt;
* Rita Johnston, University of North Carolina at Charlotte&lt;br /&gt;
* Ranti Junus, Michigan State University&lt;br /&gt;
* Martha Kyrillidou, Association of Research Libraries&lt;br /&gt;
* Eric Stedfeld, New York University&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Next Steps ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
The DLF AIG Analytics working group will present at the DLF Forum 2015 session: &amp;quot;Collaborative Efforts to Develop Best Practices in Assessment: A Progress Report&amp;quot; on Monday, October 26 at 1:30pm Pacific Time.  The session will be [http://www.diglib.org/forums/2015forum/livestream-schedule available via livestreaming] during the conference.     There will also be an opportunity to learn more about and discuss the group&#039;s work at the DLF Assessment lunch on Tuesday October 27.  &lt;br /&gt;
&lt;br /&gt;
As of the publishing of the white paper, the future of the DLF analytics working group is uncertain.   The group is looking to the digital library community for feedback, ideas, and volunteers in order to continue: should the analytics work continue? Are there other assessment efforts DLF AIG should prioritize? Please contribute to the discussion through the [https://groups.google.com/forum/#!forum/digital-library-assessment DLF AIG Google Group] or in person at DLF.   &lt;br /&gt;
&lt;br /&gt;
For those interested in continuing the analytics effort, attend the DLF sessions listed above, post to the DLF AIG Google Group or contact Molly Bragg (molly.bragg at duke.edu) and Joyce Chapman (joyce.chapman at duke.edu) directly.&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=256</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=256"/>
		<updated>2015-10-19T15:06:40Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;In 2014, Joyce Chapman, Assessment Coordinator for Duke Universities, developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which will be used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project:  &lt;br /&gt;
&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form]. The Digitization Cost Calculator 2.0 cannot be built until at least one set of data has been contributed in each of the 20 categories listed in the guidelines.&lt;br /&gt;
&lt;br /&gt;
See the cost assessment bibliography produced by the Cost Assessment working group here:&lt;br /&gt;
* [https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
&lt;br /&gt;
==Members of the Cost Assessment Working Group==&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Reissmeyer, University of California Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
==Next steps==&lt;br /&gt;
In the 2015-2016 year, the group hopes to solicit enough data contributions to build a functional digitization cost calculator 2.0. This means we need at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. &lt;br /&gt;
&lt;br /&gt;
Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here]!&lt;br /&gt;
&lt;br /&gt;
If you are interested in helping us develop best practices and guidelines for assessing costs of digital libraries, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://wiki.diglib.org/Assessment Back to the Assessment home page]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=255</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=255"/>
		<updated>2015-10-19T15:00:00Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;In 2014, Joyce Chapman, Assessment Coordinator for Duke Universities, developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was founded. This group was tasked with drafting [https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes], which will be used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project:  &lt;br /&gt;
&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
**Condition review&lt;br /&gt;
**Disbinding&lt;br /&gt;
**Fastener removal&lt;br /&gt;
**Flattening&lt;br /&gt;
**Rights review&lt;br /&gt;
**Sorting materials into items&lt;br /&gt;
**Supporting&lt;br /&gt;
**Unique identifier assignment&lt;br /&gt;
* Image capture&lt;br /&gt;
**Film or transparency scanner &lt;br /&gt;
**Flatbed scanner&lt;br /&gt;
**Manual DSLR camera&lt;br /&gt;
**Medium format camera&lt;br /&gt;
**Overhead scanner &lt;br /&gt;
**Sheet feeding scanner&lt;br /&gt;
* Quality control&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
**Level 1, 2, and 3&lt;br /&gt;
* Post-processing&lt;br /&gt;
**Alignment/rotation&lt;br /&gt;
**Background removal&lt;br /&gt;
**Clean up / dust removal&lt;br /&gt;
**Color correction and tonal adjustments&lt;br /&gt;
**Cropping images&lt;br /&gt;
**Stitching&lt;br /&gt;
* Post-preparation&lt;br /&gt;
**De-sorting&lt;br /&gt;
**Re-binding&lt;br /&gt;
**Re-fastening&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form]. The Digitization Cost Calculator 2.0 cannot be built until at least one set of data has been contributed in each of the 20 categories listed in the guidelines.&lt;br /&gt;
&lt;br /&gt;
See the cost assessment bibliography produced by the Cost Assessment working group here:&lt;br /&gt;
* [https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
&lt;br /&gt;
==Members of the Cost Assessment Working Group==&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Reissmeyer, University of California Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;br /&gt;
&lt;br /&gt;
==Next steps==&lt;br /&gt;
In the 2015-2016 year, the group hopes to solicit enough data contributions to build a functional digitization cost calculator 2.0. This means we need at least one set of data in each of the 20 categories, as well as for each type of image capture device and for each level of metadata creation and quality control. &lt;br /&gt;
&lt;br /&gt;
Help us by [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 submitting data here]!&lt;br /&gt;
&lt;br /&gt;
If you are interested in helping us develop best practices and guidelines for assessing costs of digital libraries, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up!&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=238</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=238"/>
		<updated>2015-10-15T14:58:48Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== What&#039;s the best way to assess costs for digital libraries? ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In 2014, Joyce Chapman, Assessment Coordinator for Duke Universities, developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was founded. This group was tasked with drafting [[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes]], which will be used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project:  &lt;br /&gt;
&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
* Image capture&lt;br /&gt;
* Quality control&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
* Post-processing&lt;br /&gt;
* Post-preparation&lt;br /&gt;
&lt;br /&gt;
The guidelines include definitions for a total of 20 specific data fields within these six areas, ranging from fastener removal to condition review and alignment/rotation of images.&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form]. The Digitization Cost Calculator 2.0 cannot be built until at least one set of data has been contributed in each of the 20 categories listed in the guidelines.&lt;br /&gt;
&lt;br /&gt;
If you are interested in helping us develop best practices and guidelines for assessing costs of digital libraries, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up! &lt;br /&gt;
&lt;br /&gt;
See the cost assessment bibliography produced by the Cost Assessment working group here:&lt;br /&gt;
* [https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;br /&gt;
&lt;br /&gt;
===Members of the Cost Assessment Working Group===&lt;br /&gt;
* Joyce Chapman, Duke University, (coordinator)&lt;br /&gt;
* Kinza Masood, University of Utah&lt;br /&gt;
* Chrissy Reissmeyer, University of California Santa Barbara&lt;br /&gt;
* Dan Zellner, Northwestern University&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=211</id>
		<title>Assessment:Costs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:Costs&amp;diff=211"/>
		<updated>2015-10-04T23:30:50Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== What&#039;s the best way to assess costs for digital libraries? ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
In 2014, Joyce Chapman, Assessment Coordinator for Duke Universities, developed a beta version of a [http://statelibrarync.org/plstats/digitization_calculator.php Library Digitization Cost Calculator] to assist organizations in estimating the costs for digitizing collections and initiate a conversation at the 2014 DLF Forum around cost assessment in digital libraries. In spring of 2014, the Cost Assessment working group of the DLF Assessment Interest Group was founded. This group was tasked with drafting [[https://docs.google.com/document/d/17jTJmCzKsa83BMdlgKj239Shqcbglq7I4EfxJIWWDQo/edit best practices and guidelines for the collection of time data for various digitization processes]], which will be used to standardize collection of such data in the field as well as to guide data submissions to the Digital Cost Calculator. The guidelines and definitions are organized according to some of the typical stages of a library project:  &lt;br /&gt;
&lt;br /&gt;
* Preparation of original materials&lt;br /&gt;
* Image capture&lt;br /&gt;
* Quality control&lt;br /&gt;
* Descriptive metadata creation&lt;br /&gt;
* Post-processing&lt;br /&gt;
* Post-preparation&lt;br /&gt;
&lt;br /&gt;
The guidelines include definitions for a total of 20 specific data fields within these six areas, ranging from fastener removal to condition review and alignment/rotation of images.&lt;br /&gt;
&lt;br /&gt;
In August 2015, the group produced [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wire frames] showing the planned interface and functionality of the Digitization Cost Calculator 2.0, as well as [https://drive.google.com/drive/folders/0B4veUYJxzAmSfmpMSE1zVTIxYmtzb1pFeWtiZzNFREp3bFJaYW5rYkJTV1RGWGY4Mm05LWc wireframes of how results would appear]. &lt;br /&gt;
&lt;br /&gt;
In September, the group released a [https://docs.google.com/document/d/1s1bHzkB3SSyufaoZS0JZhbVQ1AZfRgUsZ0iOcKZ3Q0k/edit call for data submissions] using the new guidelines and a new [https://duke.qualtrics.com/SE/?SID=SV_3OtqSEAbpl2QDl3 data submission form]. The Digitization Cost Calculator 2.0 cannot be built until at least one set of data has been contributed in each of the 20 categories listed in the guidelines.&lt;br /&gt;
&lt;br /&gt;
If you are interested in helping us develop best practices and guidelines for assessing costs of digital libraries, please join our &lt;br /&gt;
[https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group] and speak up! &lt;br /&gt;
&lt;br /&gt;
See the cost assessment bibliography produced by the Cost Assessment working group here:&lt;br /&gt;
* [https://docs.google.com/document/d/1yFWNWonZXHEFagI0A4ujeCl6U7SriJbfmqAgFGLA-Os/edit?usp=sharing Cost Assessment Bibliography]&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment:User_Needs&amp;diff=210</id>
		<title>Assessment:User Needs</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment:User_Needs&amp;diff=210"/>
		<updated>2015-10-04T23:03:09Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
== What are the best methods for assessing user needs with regards to digital libraries? ==&lt;br /&gt;
&lt;br /&gt;
The User and User Studies working group will be compiling resources, conducting research, and drafting literature to assist those who are interesting in evaluating users of digital repositories and their needs.&lt;br /&gt;
&lt;br /&gt;
== Addressing users and user needs ==&lt;br /&gt;
&lt;br /&gt;
Participants identified four assessment topics to explore during the [https://docs.google.com/document/d/1rLdMSPCbR_F-hE0ZhWzS70IBGEMh1450ff0x_ioPrIk/edit Assessment Breakout Session] at the 2014 DLF Forum:&lt;br /&gt;
&lt;br /&gt;
# User/Usability studies&lt;br /&gt;
# Identifying users and uses&lt;br /&gt;
# Return On Investment (ROI)&lt;br /&gt;
# Reuse of content/what do users do with content&lt;br /&gt;
&lt;br /&gt;
These four groups will make up the focus of a white paper on assessing users and user needs.  The sub-group will complete a series of tasks to generate the white paper:&lt;br /&gt;
&lt;br /&gt;
# Compile a [https://docs.google.com/document/d/1AVlT3QlmhbcUmoRtZ0hpvEhpbD9_Eld5bHsa-JNHX24/edit?usp=sharing bibliography]&lt;br /&gt;
# Develop a literature review -- completed&lt;br /&gt;
# Draft and publish a white paper on assessing users and user needs in digital libraries&lt;br /&gt;
&lt;br /&gt;
The sub-group will publish the white paper prior to the 2015 DLF Forum. Volunteers will be completing tasks in the order listed above to build up to the publication of the white paper.  If you are interested in participating in one of the tasks, please fill out our [https://docs.google.com/spreadsheets/d/16X0B-NSulh1V6HuqkSZZq0El35AQlZLbpWUlzNx2Puo/edit User Studies Participant] spreadsheet.&lt;br /&gt;
&lt;br /&gt;
Future efforts from the group may include generating complementary content, such as:&lt;br /&gt;
# Templates/guides &lt;br /&gt;
# Toolkits&lt;br /&gt;
# Original Research&lt;br /&gt;
&lt;br /&gt;
== Learning more and getting involved ==&lt;br /&gt;
&lt;br /&gt;
The sub-group maintains a [https://drive.google.com/folderview?id=0BxKlOQm6nnVndnJrR200UFZoSm8&amp;amp;usp=sharing Google Drive folder], which contains meeting notes and drafts of working documents.&lt;br /&gt;
&lt;br /&gt;
If you have any questions, or are interested in participating in the sub-group, please contact [mailto:sathompson3@uh.edu Santi Thompson].&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
	<entry>
		<id>https://wiki.diglib.org/index.php?title=Assessment&amp;diff=209</id>
		<title>Assessment</title>
		<link rel="alternate" type="text/html" href="https://wiki.diglib.org/index.php?title=Assessment&amp;diff=209"/>
		<updated>2015-09-30T15:56:50Z</updated>

		<summary type="html">&lt;p&gt;Jchapman: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&#039;&#039;&#039;Digital Library Assessment&#039;&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
Research and cultural heritage institutions are, as a matter of course, providing online access to converted and born-digital scholarly and cultural content. As the amount of that content continues to grow, there is an increased need to strategically standardize our assessment efforts. &lt;br /&gt;
&lt;br /&gt;
The [http://www.diglib.org/groups/assessment/ DLF Assessment Interest Group], founded in 2014, seeks to engage the community in developing best practices and guidelines for various kinds of digital library assessment.&lt;br /&gt;
&lt;br /&gt;
We are concerned with:&lt;br /&gt;
* determining how to measure the impact of digital collections; &lt;br /&gt;
* developing areas of commonality and benchmarks in how we measure collections across various platforms; &lt;br /&gt;
* understanding cost and benefit of digital collections; and &lt;br /&gt;
* exploring how can we best collect, analyze, communicate, and share such information effectively across our various stakeholders—from collection managers to scholars.&lt;br /&gt;
&lt;br /&gt;
This wiki will be a central location for documentation and collection of resources to assist those seeking to assess their digital libraries.&lt;br /&gt;
&lt;br /&gt;
Using the [https://docs.google.com/document/d/18kOolh-U0qIrT3tcJg8fXMnN5pnKylFpxvtRL561baU/edit?usp=sharing Digital Library Assessment Framework], we formed four working groups in the fall of 2014 in areas of strong interest to the DLF AIG community:&lt;br /&gt;
* [[Assessment:Costs]]&lt;br /&gt;
* [[Assessment:Citations]]  &lt;br /&gt;
* [[Assessment:User Needs]]&lt;br /&gt;
* [[Assessment:Analytics]]&lt;br /&gt;
&lt;br /&gt;
Each of these four groups has been working over the past year to develop white papers and tools, which will be presented at the 2015 DLF Forum in October. &lt;br /&gt;
&lt;br /&gt;
To take part in the conversation, join the [https://groups.google.com/forum/#!forum/digital-library-assessment Digital Library Assessment Google Group], which is open to anyone interested in learning about or collaborating on the improvement of digital library assessment.&lt;/div&gt;</summary>
		<author><name>Jchapman</name></author>
	</entry>
</feed>