The document discusses data quality issues that arise from aggregation at large scales, like the Digital Public Library of America (DPLA). It notes that aggregated data is often heterogeneous, relies primarily on basic metadata rather than text, and comes from various sources without consistent standards. This can lead to technical problems like a lack of normalization and content problems like meaningless, missing, confusing or incomplete values. It proposes several initiatives could help address these issues, such as analyzing data quality in the DPLA workflow, reviewing mandatory elements, and taking inspiration from Europeana's Data Quality Committee.