<?xml version="1.0" encoding="UTF-8" ?><!-- generator=Zoho Sites --><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><atom:link href="https://www.nownextlater.ai/Insights/tag/dataset-development/feed" rel="self" type="application/rss+xml"/><title>Now Next Later AI - Blog #Dataset development</title><description>Now Next Later AI - Blog #Dataset development</description><link>https://www.nownextlater.ai/Insights/tag/dataset-development</link><lastBuildDate>Wed, 26 Nov 2025 21:38:21 +1100</lastBuildDate><generator>http://zoho.com/sites/</generator><item><title><![CDATA[The Emerging Task of Measuring AI Training Data]]></title><link>https://www.nownextlater.ai/Insights/post/the-emerging-task-of-measuring-ai-training-data</link><description><![CDATA[A new perspective paper argues for "measuring data" as a critical task to advance responsible AI development. Just as physical objects can be measured, data used to train AI systems should also be quantitatively analyzed to understand its composition.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_ZvHxJFMoSq-98yCwxYPHgA" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_tZt48zXQTsKXT1f1oGLSlw" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_MoNhZsgtRQqy0iGLR8-9XQ" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_EL9akuHOsBh9OiTwHe18IA" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_EL9akuHOsBh9OiTwHe18IA"] .zpimage-container figure img { width: 500px ; height: 399.67px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_EL9akuHOsBh9OiTwHe18IA"] .zpimage-container figure img { width:500px ; height:399.67px ; } } @media (max-width: 767px) { [data-element-id="elm_EL9akuHOsBh9OiTwHe18IA"] .zpimage-container figure img { width:500px ; height:399.67px ; } } [data-element-id="elm_EL9akuHOsBh9OiTwHe18IA"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-13%20at%208.11.45%20pm.png" width="500" height="399.67" loading="lazy" size="medium" alt="Examples of different data measurements proposed in image- and language-based data science and machine learning, alongside analogs in the physical sciences." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_3kksYDX2SDSTKy-adgtOWQ" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_3kksYDX2SDSTKy-adgtOWQ"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>A new perspective paper argues for &quot;measuring data&quot; as a critical task to advance responsible AI development. Just as physical objects can be measured, data used to train AI systems should also be quantitatively analyzed to understand its composition.</p><p><br></p><p>The authors propose formalizing data measurement as a core research area. They categorize different measurement types like distance, density, and diversity. Distance metrics like Euclidean distance capture similarities in the data. Density reflects how densely certain concepts are represented. <span style="font-family:&quot;Questrial&quot;, sans-serif;">Diversity measures heterogeneity.</span></p><p><br></p><p>The paper summarizes measurements proposed across computer vision, natural language processing, and general data science. For instance, perplexity scores indicate how predictable text data is. Image density models visual patterns. Word mover's distance compares document similarity based on word embeddings.</p><p><br></p><p>The authors argue precise data measurements will let practitioners better curate datasets and understand what models learn. Characterizing training data is crucial for responsible AI. Metrics can reveal underrepresentation of groups or embed problematic biases.</p><p><br></p><p>For business leaders deploying AI, this paper provides a framework to evaluate data rigorously before training models. Combining multiple measurements will enable deeply understanding datasets. Just as standardized physical measurements enabled advances in science, quantifying data properties will drive progress in AI. Insisting models be trained on measured data is vital for ethical AI.</p><p><br></p><p>Sources:</p><div style="color:inherit;"><div><div><div><div><p><a href="https://www.researchgate.net/publication/366212996_Measuring_Data" title="Measuring Data " rel="">Measuring Data </a></p><p></p></div>
</div></div></div></div><p></p></div><p></p></div></div></div></div></div></div></div>
 ]]></content:encoded><pubDate>Sun, 13 Aug 2023 20:14:16 +1000</pubDate></item><item><title><![CDATA[Making Data Work More Visible Through Documentation]]></title><link>https://www.nownextlater.ai/Insights/post/making-data-work-more-visible-through-documentation</link><description><![CDATA[A new study provides insights into the complex processes and people behind ML data work.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_TwxavckZQA29ypYTg48xKQ" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_7iuWKqXDRi-wZLEV3mpYEA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_ri8OmA1yRMSzjyala3dkJA" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_JLlVHh2c5kXu79e-4LkG7Q" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_JLlVHh2c5kXu79e-4LkG7Q"] .zpimage-container figure img { width: 800px ; height: 492.50px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_JLlVHh2c5kXu79e-4LkG7Q"] .zpimage-container figure img { width:500px ; height:307.81px ; } } @media (max-width: 767px) { [data-element-id="elm_JLlVHh2c5kXu79e-4LkG7Q"] .zpimage-container figure img { width:500px ; height:307.81px ; } } [data-element-id="elm_JLlVHh2c5kXu79e-4LkG7Q"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-large zpimage-tablet-fallback-large zpimage-mobile-fallback-large hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-13%20at%2012.29.34%20pm.png" width="500" height="307.81" loading="lazy" size="large" alt="Documenting Data Production Processes: A Participatory Approach for Data Work" data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_c4Hd3kcRQaWRmKFCcKoDSA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_c4Hd3kcRQaWRmKFCcKoDSA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Machine learning (ML) systems are becoming increasingly prevalent, but few business leaders understand how the data used to train these systems is produced. A new study provides insights into the complex processes and people behind ML data work.</p><p><br></p><p>The study was conducted by researchers who partnered with two business process outsourcing (BPO) companies that employ human workers to collect and label data used for training machine learning systems. Through interviews and design workshops with data workers, managers, and clients, the researchers aimed to understand how documentation practices could be improved to capture the context and processes involved in producing ML training data.</p><p><br></p><p>Key findings:</p><ul><li>Data workers have limited visibility into the overall goals and uses of the projects they work on. Better documentation that explains the client, project timeline, and intended uses of the final dataset could improve coordination and workers’ sense of meaning.</li><li>Data work involves collaboration between many geographically dispersed actors, but communication tends to be top-down. Enabling feedback loops in documentation could foster more participatory, iterative processes between clients, managers, and data workers.</li><li>Data work is dynamic, but documentation practices are often static. Tools like revision histories and timeline tracking could better reflect the evolving nature of projects and datasets.</li><li>Worker needs are not fully represented in current documentation. Including details like task instructions, payment terms, and ethical considerations could make documentation more worker-centric.</li></ul><p><br></p><p>The researchers conclude that &quot;data production documentation should travel across multiple stakeholders and facilitate their communication.&quot; Viewing documentation as a &quot;boundary object&quot; that adapts to diverse needs while maintaining integrity could enable more participatory and equitable data production processes.</p><p><br></p><p>Key implications for business leaders:</p><ul><li>Seek more visibility into how your ML training data is produced and provide input to vendors on documentation needs.</li><li>Recognize and engage data workers as key collaborators, not just low-level laborers. Their feedback can improve dataset quality.</li><li>Set documentation policies that make values like transparency, participation, and fairness explicit. This also mitigates business risks.</li><li>Documentation supports coordination and accountability. Invest in tools and practices that reflect the dynamic reality of data work.</li></ul><div><br><span style="color:inherit;"></span></div><div><div><span style="color:inherit;font-size:14px;"><span style="font-weight:400;text-indent:0px;">Machine Learning's pervasive growth underscores the importance of understanding its foundational data. This study underscores that the intricate human processes behind data collection and labeling are often underappreciated and misunderstood. As businesses increasingly rely on ML systems, there's a compelling need for enriched documentation that facilitates comprehensive communication across stakeholders, recognizes the invaluable contribution of data workers, and remains dynamic to the evolving nature of data projects. By viewing documentation as an adaptable &quot;boundary object,&quot; businesses can champion a more inclusive, transparent, and effective ML data production process, ensuring both operational excellence and ethical integrity.</span></span></div></div><div><span style="color:inherit;"><span style="font-size:16px;font-weight:400;text-indent:0px;"><br></span></span></div><div>Sources:</div><div><div><span style="color:inherit;"><a href="https://arxiv.org/abs/2207.04958" title="Documenting Data Production Processes: A Participatory Approach for Data Work" rel="">Documenting Data Production Processes: A Participatory Approach for Data Work</a><br>Milagros Miceli, Tianling Yang, Adriana Alvarado Garcia, Julian Posada, Sonja Mei Wang, Marc Pohl, Alex Hanna</span></div></div></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Sun, 13 Aug 2023 12:31:31 +1000</pubDate></item><item><title><![CDATA[Examining How AI Training Datasets Are Built: A Framework for More Responsible Practices]]></title><link>https://www.nownextlater.ai/Insights/post/responsible-ai-starts-with-understanding-dataset-development</link><description><![CDATA[In a recent paper, researchers Mehtab Khan and Alex Hanna highlight the need for greater scrutiny, transparency, and accountability in how massive datasets for machine learning models are created.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_BrJ1XTJ-QCynLbFRRN4wjg" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_ICxg0lAsTzqAGytnGKaiPA" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_U_414SkTTsmJ7snzakDfiw" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_ubaW648SWMcPGPAG6CbEAQ" data-element-type="image" class="zpelement zpelem-image "><style> @media (min-width: 992px) { [data-element-id="elm_ubaW648SWMcPGPAG6CbEAQ"] .zpimage-container figure img { width: 500px ; height: 370.48px ; } } @media (max-width: 991px) and (min-width: 768px) { [data-element-id="elm_ubaW648SWMcPGPAG6CbEAQ"] .zpimage-container figure img { width:500px ; height:370.48px ; } } @media (max-width: 767px) { [data-element-id="elm_ubaW648SWMcPGPAG6CbEAQ"] .zpimage-container figure img { width:500px ; height:370.48px ; } } [data-element-id="elm_ubaW648SWMcPGPAG6CbEAQ"].zpelem-image { border-radius:1px; } </style><div data-caption-color="" data-size-tablet="" data-size-mobile="" data-align="center" data-tablet-image-separate="false" data-mobile-image-separate="false" class="zpimage-container zpimage-align-center zpimage-size-medium zpimage-tablet-fallback-medium zpimage-mobile-fallback-medium hb-lightbox " data-lightbox-options="
                type:fullscreen,
                theme:dark"><figure role="none" class="zpimage-data-ref"><span class="zpimage-anchor" role="link" tabindex="0" aria-label="Open Lightbox" style="cursor:pointer;"><picture><img class="zpimage zpimage-style-none zpimage-space-none " src="/Screenshot%202023-08-13%20at%2011.48.34%20am.png" width="500" height="370.48" loading="lazy" size="medium" alt="A breakdown of which group is susceptible to harm at every stage of the AI dataset and model development stage." data-lightbox="true"/></picture></span></figure></div>
</div><div data-element-id="elm_JnFD1VaJToWpwjPIS27WwA" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_JnFD1VaJToWpwjPIS27WwA"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><div style="color:inherit;"><div style="color:inherit;"><p>As artificial intelligence (AI) and machine learning technologies become more widely adopted across sectors, businesses must prioritize responsible and ethical implementation. A critical yet often overlooked area is how AI training datasets are constructed behind the scenes.</p><p><br></p><p>In a recent paper, researchers Mehtab Khan and Alex Hanna highlight the need for greater scrutiny, transparency, and accountability in how massive datasets for machine learning models are created. Their analysis proposes breaking the process into distinct stages and identifying affected individuals.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">Legally Gray Areas Around Data Collection</span></p><p><br></p><p>The paper notes that current copyright law is ambiguous regarding reproducing large volumes of text, images, and other data for assembling AI training datasets through scraping websites and other sources. While fair use exemptions may technically apply in some cases, the legal boundaries are highly complex. This gray area allows widespread copying to build datasets, but concerning practices could still raise issues like copyright infringement.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">Emerging Individual and Collective Privacy Concerns</span></p><p><br></p><p>Collecting data en masse, even from publicly available websites, risks violating the privacy of individuals included without their consent. But the paper emphasizes privacy issues extend beyond individual data collection. Biases and unfair representations affecting certain populations also clearly emerge when examining dataset design choices.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">A Framework for the Entire Development Pipeline</span></p><p><br></p><p>To illuminate interconnected problems throughout the process, the researchers propose systematically analyzing each stage of assembling training datasets. This includes initial problem definition, data collection and cleaning, annotating examples, model building and evaluation, and distribution. Examining how different stakeholders are impacted at each phase can uncover ethical issues.</p><p><br></p><p>The paper argues this framework highlights where greater transparency, oversight, and accountability are most needed. It can also inform policies and self-regulation. Documenting datasets is one step, but comprehensive governance of the full development lifecycle is required for responsible AI.</p><p><br></p><p><span style="font-family:&quot;Oswald&quot;, sans-serif;">Proactive Policies Needed</span></p><p><br></p><p>Understanding training data origins is no simple task but crucial for mitigating risks like biases, discrimination, and improper data usage. While technical teams focus on accuracy, leaders must invest in responsible data sourcing and stewardship. Establishing proactive policies and drawing on frameworks like the one outlined in this paper can lead to more ethical AI systems.</p><p><br></p><p>Sources:</p><p><span style="color:inherit;"><a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4217148" title="The Subjects and Stages of AI Dataset Development: A Framework for Dataset Accountability" rel="">The Subjects and Stages of AI Dataset Development: A Framework for Dataset Accountability</a></span></p><p></p><div style="color:inherit;"><div><div><div><p>Mehtab Khan, Alex Hanna</p></div>
</div></div></div><p></p></div></div></div><p></p></div></div></div></div></div></div>
</div> ]]></content:encoded><pubDate>Sun, 13 Aug 2023 11:51:01 +1000</pubDate></item><item><title><![CDATA[Documenting Machine Learning Datasets to Increase Accountability and Inclusivity]]></title><link>https://www.nownextlater.ai/Insights/post/documenting-machine-learning-datasets-to-increase-accountability-and-inclusivity</link><description><![CDATA[Machine learning models rely heavily on their training datasets, inheriting inherent biases and limitations. This research proposes "datasheets for datasets" increasing transparency and mitigating risks.]]></description><content:encoded><![CDATA[<div class="zpcontent-container blogpost-container "><div data-element-id="elm_hdGjxQLpRvyyJP2Az_eb2g" data-element-type="section" class="zpsection "><style type="text/css"></style><div class="zpcontainer-fluid zpcontainer"><div data-element-id="elm_tjAkd_ubTX2QsxUNfgdPwQ" data-element-type="row" class="zprow zprow-container zpalign-items- zpjustify-content- " data-equal-column=""><style type="text/css"></style><div data-element-id="elm_XhYkB4p6SlKLQ7Yu6XRaWQ" data-element-type="column" class="zpelem-col zpcol-12 zpcol-md-12 zpcol-sm-12 zpalign-self- "><style type="text/css"></style><div data-element-id="elm_yOi6ysW-SeK00iw4ZjhFuw" data-element-type="text" class="zpelement zpelem-text "><style> [data-element-id="elm_yOi6ysW-SeK00iw4ZjhFuw"].zpelem-text { border-radius:1px; } </style><div class="zptext zptext-align-left " data-editor="true"><div style="color:inherit;"><p>Machine learning models rely heavily on their training datasets, inheriting inherent biases and limitations. Yet unlike other fields like electronics where datasheets meticulously detail components' operating parameters, no standards exist for documenting machine learning datasets. This research proposes &quot;datasheets for datasets&quot; to fill that gap, increasing transparency and mitigating risks.</p><p><br></p><p>Datasets impart significant yet often opaque influences on models. Biased data produces biased AI, like résumé screening tools that disadvantaged women. Undocumented datasets also limit reproducibility. The World Economic Forum thus recommends documenting datasets to avoid discrimination.</p><p>The authors devised an initial set of questions for dataset creators to reflect on a dataset's motivation, composition, collection, processing, uses, distribution and maintenance. For example, what subpopulations does the data represent? Could it directly or indirectly identify individuals? Might it result in unfair treatment for certain groups?</p><p><br></p><p>They refined the questions based on creating sample datasheets and feedback from companies pilots. A legal review led to removing explicit regulatory compliance questions. The goal is to encourage thoughtful dataset creation versus check-box documentation.</p><p><br></p><p>Published datasheets have already enabled better model development. At Microsoft, reviewing datasheets surfaced imbalanced gender data. But adding synthetic balanced data produced more equitable models. Datasheets also helped researchers select appropriate datasets.</p><p><br></p><p>Challenges remain in adapting datasheets across domains like dynamic datasets. Implementation imposes overhead on creators. And datasheets alone cannot identify every dataset bias or misuse. Combining with expertise like social scientists is recommended when data involves people.</p><p><br></p><p>Still, datasheets facilitate communication between creators and users, increasing intentionality in dataset development. Distinguishing datasets via transparency aids accountability and quality. For businesses, documenting training data provides reputational benefits and ensures models reflect corporate values. Proactively auditing algorithms and data is crucial as AI proliferates in customer-facing systems.</p><p><br></p><p>More broadly, the paper argues datasheets exemplify responsible data science in service of justice. Much like nutrition labels detail food ingredients for healthier eating, datasheets elucidate datasets' impacts. The authors urge centering inclusion and examining unintended consequences in design.</p><p>Responsible innovation requires transcending technical fixes to ask philosophical questions. What societal outcomes do we want these technologies to help achieve? How can AI empower communities historically excluded by “neutral” systems? Keeping diverse perspectives at the heart of development is key.</p><p><br></p><p>While datasheets provide a practical tool for dataset creators and consumers, their significance is as much symbolic as practical. They represent taking an intentional stance on transparency and equity in an opaque, high-stakes field. For businesses navigating AI's rise, this study offers inspiration to wield these technologies thoughtfully for human progress.</p><p><br></p><p>Sources:</p><p><span style="color:inherit;"><a href="https://dl.acm.org/doi/10.1145/3458723" title="Datasheets for datasets" rel="">Datasheets for datasets</a>, </span><span style="color:inherit;"><span>Timnit Gebru, </span></span><span style="color:inherit;"><span>Jamie Morgenstern, </span></span><span style="color:inherit;"><span>Briana Vecchione, </span></span><span style="color:inherit;"><span>Jennifer Wortman Vaughan, </span></span><span style="color:inherit;"><span>Hanna Wallach, </span></span><span style="color:inherit;"><span>Hal Daumé III,</span></span><span style="color:inherit;"><span> Kate Crawford</span></span></p><p></p><p></p></div><p></p></div>
</div></div></div></div></div></div> ]]></content:encoded><pubDate>Sun, 13 Aug 2023 11:21:07 +1000</pubDate></item></channel></rss>