Category Archives: Machine Learning
A causal perspective on dataset bias in machine learning for medical imaging – Nature.com
Char, D. S., Shah, N. H. & Magnus, D. Implementing machine learning in health care addressing ethical challenges. N. Engl. J. Med. 378, 981983 (2018).
Article PubMed PubMed Central Google Scholar
Obermeyer, Z., Powers, B., Vogeli, C. & Mullainathan, S. Dissecting racial bias in an algorithm used to manage the health of populations. Science 366, 447453 (2019).
Article ADS CAS PubMed Google Scholar
Wiens, J. et al. Do no harm: a roadmap for responsible machine learning for health care. Nat. Med. 25, 13371340 (2019).
Article CAS PubMed Google Scholar
Buolamwini, J. & Gebru, T. Gender shades: intersectional accuracy disparities in commercial gender classification. In Proc. 1st Conference on Fairness, Accountability and Transparency (eds Friedler, S. A. & Wilson, C.) 7791 (PMLR, 2018).
Beede, E. et al. A human-centered evaluation of a deep learning system deployed in clinics for the detection of diabetic retinopathy. In Proc. 2020 CHI Conference on Human Factors in Computing Systems 112 (Association for Computing Machinery, 2020).
Seyyed-Kalantari, L., Liu, G., McDermott, M., Chen, I. Y. & Ghassemi, M. CheXclusion: fairness gaps in deep chest X-ray classifiers. Pacific Symp. Biocomput. 26, 232243 (World Scientific, 2021).
Seyyed-Kalantari, L., Zhang, H., McDermott, M. B., Chen, I. Y. & Ghassemi, M. Underdiagnosis bias of artificial intelligence algorithms applied to chest radiographs in under-served patient populations. Nat. Med. 27, 21762182 (2021).
Article CAS PubMed PubMed Central Google Scholar
Mamary, A. J. et al. Race and gender disparities are evident in COPD underdiagnoses across all severities of measured airflow obstruction. Chronic Obstruct. Pulmon. Dis. 5, 177184 (2018).
Google Scholar
Oakden-Rayner, L., Dunnmon, J., Carneiro, G. & R, C. Hidden stratification causes clinically meaningful failures in machine learning for medical imaging. Proc. ACM Conf. Health Infer. Learn. 2020, 151159 (2020).
Article Google Scholar
Gianfrancesco, M. A., Tamang, S., Yazdany, J. & Schmajuk, G. Potential biases in machine learning algorithms using electronic health record data. JAMA Intern. Med. 178, 15441547 (2018).
Article PubMed PubMed Central Google Scholar
Larrazabal, A. J., Nieto, N., Peterson, V., Milone, D. H. & Ferrante, E. Gender imbalance in medical imaging datasets produces biased classifiers for computer-aided diagnosis. Proc. Natl Acad. Sci. USA 117, 1259212594 (2020).
Article ADS CAS PubMed PubMed Central Google Scholar
Wang, Z. et al. Towards fairness in visual recognition: effective strategies for bias mitigation. In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition 89168925 (IEEE, 2020).
Zietlow, D. et al. Leveling down in computer vision: pareto inefficiencies in fair deep classifiers. In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition 1041010421 (IEEE, 2022).
Alvi, M., Zisserman, A. & Nellaaker, C. Turning a blind eye: explicit removal of biases and variation from deep neural network embeddings. In Proc. European Conference on Computer Vision Workshops 556572 (Springer, 2018).
Kim, B., Kim, H., Kim, K., Kim, S. & Kim, J. Learning not to learn: training deep neural networks with biased data. In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition 90129020 (IEEE, 2019).
Madras, D., Creager, E., Pitassi, T. & Zemel, R. Learning adversarially fair and transferable representations. In International Conference on Machine Learning 33843393 (PMLR, 2018).
Edwards, H. & Storkey, A. Censoring representations with an adversary. In International Conference in Learning Representations (eds Bengio, Y. & LeCun, Y.) (2016). Editors: Yoshua Bengio and Yann LeCun.
Ramaswamy, V. V., Kim, S. S. Y. & Russakovsky, O. Fair attribute classification through latent space de-biasing. In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition 93019310 (IEEE, 2021).
Wang, M., Deng, W., Hu, J., Tao, X. & Huang, Y. Racial faces in the wild: reducing racial bias by information maximization adaptation network. In Proc. IEEE/CVF International Conference on Computer Vision 692702 (IEEE, 2019).
Hendricks, L. A., Burns, K., Saenko, K., Darrell, T. & Rohrbach, A. Women also snowboard: overcoming bias in captioning models. In Computer Vision ECCV 2018 Vol. 11207 (eds Ferrari, V. et al.) 793811 (Springer, 2018).
Li, Y. & Vasconcelos, N. REPAIR: removing representation bias by dataset resampling. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition 95649573 (IEEE, 2019).
Quadrianto, N., Sharmanska, V. & Thomas, O. Discovering fair representations in the data domain. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition 82198228 (IEEE, 2019).
Wang, T., Zhao, J., Yatskar, M., Chang, K.-W. & Ordonez, V. Balanced datasets are not enough: estimating and mitigating gender bias in deep image representations. In 2019 IEEE/CVF International Conference on Computer Vision 53095318 (IEEE, 2019).
Corbett-Davies, S. & Goel, S. The measure and mismeasure of fairness: a critical review of fair machine learning. Preprint at https://arxiv.org/abs/1808.00023 (2018).
Friedler, S. A. et al. A comparative study of fairness-enhancing interventions in machine learning. In Proc. Conference on Fairness, Accountability, and Transparency 329338 (Association for Computing Machinery, 2019).
Zong, Y., Yang, Y. & Hospedales, T. MEDFAIR: benchmarking fairness for medical imaging. In International Conference on Learning Representations (eds Kim, B., Nickel, M., Wang, M., Chen, N. F. & Marivate, V.) (2023).
Castro, D. C., Walker, I. & Glocker, B. Causality matters in medical imaging. Nat. Commun. 11, 3673 (2020).
Article ADS CAS PubMed PubMed Central Google Scholar
Subbaswamy, A. & Saria, S. From development to deployment: dataset shift, causality, and shift-stable models in health AI. Biostatistics 21, 345352 (2020).
MathSciNet PubMed Google Scholar
Subbaswamy, A. & Saria, S. Counterfactual normalization: proactively addressing dataset shift using causal mechanisms. In 34th Conference on Uncertainty in Artificial Intelligence 2018 947957 (Association For Uncertainty in Artificial Intelligence, 2018).
Subbaswamy, A., Schulam, P. & Saria, S. Preventing failures due to dataset shift: learning predictive models that transport. In Proc. Twenty-Second International Conference on Artificial Intelligence and Statistics 31183127 (PMLR, 2019).
Huang, B. et al. Behind distribution shift: mining driving forces of changes and causal arrows. Proc. IEEE Int. Conf. Data Mining 2017, 913918 (2017).
Google Scholar
Yue, Z., Sun, Q., Hua, X.-S. & Zhang, H. Transporting causal mechanisms for unsupervised domain adaptation. In Proc. IEEE/CVF International Conference on Computer Vision 2021 85998608 (IEEE, 2021).
Zhang, K., Gong, M. & Schoelkopf, B. Multi-source domain adaptation: a causal view. In Proceedings of the AAAI Conference on Artificial Intelligence 29, 31503157 (AAAI Press, Palo Alto, CA, 2015).
Magliacane, S. et al. Domain adaptation by using causal inference to predict invariant conditional distributions. In Proc. 32nd International Conference on Neural Information Processing Systems 1086910879 (Curran Associates Inc., 2018).
Chen, R. J. et al. Algorithmic fairness in artificial intelligence for medicine and healthcare. Nat. Biomed. Eng. 7, 719742 (2023).
Article PubMed PubMed Central Google Scholar
Vapnik, V. An overview of statistical learning theory. IEEE Trans. Neur. Netw. 10, 988999 (1999).
Article CAS Google Scholar
Peters, J., Janzing, D. & Schlkopf, B. Elements of Causal Inference: Foundations and Learning Algorithms (MIT Press, 2017).
Pearl, J. Causality: Models, Reasoning, and Inference 2nd edn (Cambridge Univ. Press, 2011).
Schlkopf, B. et al. On causal and anticausal learning. In Proc. 29th International Coference on Machine Learning 459466 (Omnipress, 2012).
Verma, T. & Pearl, J. Causal networks: semantics and expressiveness. In Proc. Fourth Annual Conference on Uncertainty in Artificial Intelligence 6978 (North-Holland Publishing Co., 1990).
Pearl, J. & Dechter, R. Identifying independencies in causal graphs with feedback. In Proc. Twelfth International Conference on Uncertainty in Artificial Intelligence 420426 (Morgan Kaufmann Publishers Inc., 1996).
Glocker, B., Jones, C., Bernhardt, M. & Winzeck, S. Algorithmic encoding of protected characteristics in chest X-ray disease detection models. eBioMedicine 89, 104467 (2023).
Article PubMed PubMed Central Google Scholar
Gichoya, J. W. et al. AI recognition of patient race in medical imaging: a modelling study. Lancet Digit. Health 4, e406e414 (2022).
Article CAS PubMed PubMed Central Google Scholar
Jones, C., Roschewitz, M. & Glocker, B. The role of subgroup separability in group-fair medical image classification. In Medical Image Computing and Computer Assisted Intervention 2023 179188 (Springer Nature, 2023).
Mccradden, M. et al. Whats fair is fair? Presenting JustEFAB, an ethical framework for operationalizing medical ethics and social justice in the integration of clinical machine learning: JustEFAB. In Proc. 2023 ACM Conference on Fairness, Accountability, and Transparency 15051519 (Association for Computing Machinery, 2023).
Chiappa, S. Path-specific counterfactual fairness. In Proceedings of the AAAI Conference on Artificial Intelligence 33, 78017808 (AAAI Press, Palo Alto, CA, 2019).
Friedler, S. A., Scheidegger, C. & Venkatasubramanian, S. On the (im)possibility of fairness. Preprint at https://arxiv.org/abs/1609.07236 (2016).
Wachter, S., Mittelstadt, B. & Russell, C. Bias preservation in machine learning: the legality of fairness metrics under EU non-discrimination law. West Virginia Law Review 123, 735790 (2021).
Hardt, M., Price, E. & Srebro, N. Equality of opportunity in supervised learning. In Advances in Neural Information Processing Systems (eds Lee, D. et al.) 29, 33233331 (Curran Associates, 2016).
Zemel, R., Wu, Y., Swersky, K., Pitassi, T. & Dwork, C. Learning fair representations. In Proc. 30th International Conference on Machine Learning 325333 (PMLR, 2013).
Dutta, S. et al. Is there a trade-off between fairness and accuracy? A perspective using mismatched hypothesis testing. In Proc. 37th International Conference on Machine Learning 28032813 (PMLR, 2020).
Wick, M., panda, s. & Tristan, J.-B. Unlocking Fairness: A Trade-off Revisited. In Advances in Neural Information Processing Systems Vol. 32 (Curran Associates, Inc., 2019).
Plecko, D. & Bareinboim, E. Causal fairness analysis. Preprint at https://arxiv.org/abs/2207.11385 (2022).
Mao, C. et al. Causal transportability for visual recognition. In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition 75217531 (IEEE, 2022).
Pearl, J. & Bareinboim, E. Transportability of causal and statistical relations: a formal approach. In Proceedings of the AAAI Conference on Artificial Intelligence 25, 247254 (AAAI Press, Palo Alto, CA, 2011).
Jiang, Y. & Veitch, V. Invariant and transportable representations for anti-causal domain shifts. Adv. Neur. Inf. Process. Syst. 35, 2078220794 (2022).
Google Scholar
Wolpert, D. & Macready, W. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1, 6782 (1997).
Article Google Scholar
Holland, P. W. Statistics and causal inference. J. Am. Stat. Assoc. 81, 945960 (1986).
Article MathSciNet Google Scholar
Schrouff, J. et al. Diagnosing failures of fairness transfer across distribution shift in real-world medical settings. In Advances in Neural Information Processing Systems 3, 1930419318 (Curran Associates, 2022).
Bernhardt, M., Jones, C. & Glocker, B. Potential sources of dataset bias complicate investigation of underdiagnosis by machine learning algorithms. Nat. Med. 28, 11571158 (2022).
Article CAS PubMed Google Scholar
Szczepura, A. Access to health care for ethnic minority populations. Postgrad. Med. J. 81, 141147 (2005).
Article CAS PubMed PubMed Central Google Scholar
Richardson, L. D. & Norris, M. Access to health and health care: how race and ethnicity matter. Mt Sinai J. Med. 77, 166177 (2010).
Article PubMed Google Scholar
Niccoli, T. & Partridge, L. Ageing as a risk factor for disease. Curr. Biol. 22, R741752 (2012).
Article CAS PubMed Google Scholar
Riedel, B. C., Thompson, P. M. & Brinton, R. D. Age, APOE and sex: triad of risk of Alzheimers disease. J. Steroid Biochem. Molec. Biol. 160, 134147 (2016).
Article CAS PubMed Google Scholar
Dwork, C., Immorlica, N., Kalai, A. T. & Leiserson, M. Decoupled classifiers for group-fair and efficient machine learning. In Proc. 1st Conference on Fairness, Accountability and Transparency Vol. 81 (eds Friedler, S. A. & Wilson, C.) 119133 (PMLR, 2018).
Boyko, E. J. & Alderman, B. W. The use of risk factors in medical diagnosis: opportunities and cautions. J. Clin. Epidemiol. 43, 851858 (1990).
Article CAS PubMed Google Scholar
Iglehart, J. K. Health insurers and medical-imaging policya work in progress. N. Engl. J. Med. 360, 10301037 (2009).
Article CAS PubMed Google Scholar
Iglehart, J. K. The new era of medical imagingprogress and pitfalls. N. Engl. J. Med. 354, 28222828 (2006).
Article CAS PubMed Google Scholar
Irvin, J. et al. CheXpert: a large chest radiograph dataset with uncertainty labels and expert comparison. Proc. AAAI Conf. Artif. Intell. 33, 590597 (2019).
Google Scholar
Johnson, A. E. W. et al. MIMIC-CXR, a de-identified publicly available database of chest radiographs with free-text reports. Sci. Data 6, 317 (2019).
Article PubMed PubMed Central Google Scholar
Jiang, H. & Nachum, O. Identifying and correcting label bias in machine learning. In Proc. Twenty Third International Conference on Artificial Intelligence and Statistics 702712 (PMLR, 2020).
Gebru, T. et al. Datasheets for datasets. Commun. ACM 64, 8692 (2021).
Read more:
A causal perspective on dataset bias in machine learning for medical imaging - Nature.com
MIT researchers remotely map crops, field by field – MIT News
Crop maps help scientists and policymakers track global food supplies and estimate how they might shift with climate change and growing populations. But getting accurate maps of the types of crops that are grown from farm to farm often requires on-the-ground surveys that only a handful of countries have the resources to maintain.
Now, MIT engineers have developed a method to quickly and accurately label and map crop types without requiring in-person assessments of every single farm. The teams method uses a combination of Google Street View images, machine learning, and satellite data to automatically determine the crops grown throughout a region, from one fraction of an acre to the next.
The researchers used the technique to automatically generate the first nationwide crop map of Thailand a smallholder country where small, independent farms make up the predominant form of agriculture. The team created a border-to-border map of Thailands four major crops rice, cassava, sugarcane, and maize and determined which of the four types was grown, at every 10 meters, and without gaps, across the entire country. The resulting map achieved an accuracy of 93 percent, which the researchers say is comparable to on-the-ground mapping efforts in high-income, big-farm countries.
The team is applying their mapping technique to other countries such as India, where small farms sustain most of the population but the type of crops grown from farm to farm has historically been poorly recorded.
Its a longstanding gap in knowledge about what is grown around the world, says Sherrie Wang, the dArbeloff Career Development Assistant Professor in MITs Department of Mechanical Engineering, and the Institute for Data, Systems, and Society (IDSS). The final goal is to understand agricultural outcomes like yield, and how to farm more sustainably. One of the key preliminary steps is to map what is even being grown the more granularly you can map, the more questions you can answer.
Wang, along with MIT graduate student Jordi Laguarta Soler and Thomas Friedel of the agtech company PEAT GmbH, will present a paper detailing their mapping method later this month at the AAAI Conference on Artificial Intelligence.
Ground truth
Smallholder farms are often run by a single family or farmer, who subsist on the crops and livestock that they raise. Its estimated that smallholder farms support two-thirds of the worlds rural population and produce 80 percent of the worlds food. Keeping tabs on what is grown and where is essential to tracking and forecasting food supplies around the world. But the majority of these small farms are in low to middle-income countries, where few resources are devoted to keeping track of individual farms crop types and yields.
Crop mapping efforts are mainly carried out in high-income regions such as the United States and Europe, where government agricultural agencies oversee crop surveys and send assessors to farms to label crops from field to field. These ground truth labels are then fed into machine-learning models that make connections between the ground labels of actual crops and satellite signals of the same fields. They then label and map wider swaths of farmland that assessors dont cover but that satellites automatically do.
Whats lacking in low- and middle-income countries is this ground label that we can associate with satellite signals, Laguarta Soler says. Getting these ground truths to train a model in the first place has been limited in most of the world.
The team realized that, while many developing countries do not have the resources to maintain crop surveys, they could potentially use another source of ground data: roadside imagery, captured by services such as Google Street View and Mapillary, which send cars throughout a region to take continuous 360-degree images with dashcams and rooftop cameras.
In recent years, such services have been able to access low- and middle-income countries. While the goal of these services is not specifically to capture images of crops, the MIT team saw that they could search the roadside images to identify crops.
Cropped image
In their new study, the researchers worked with Google Street View (GSV) images taken throughout Thailand a country that the service has recently imaged fairly thoroughly, and which consists predominantly of smallholder farms.
Starting with over 200,000 GSV images randomly sampled across Thailand, the team filtered out images that depicted buildings, trees, and general vegetation. About 81,000 images were crop-related. They set aside 2,000 of these, which they sent to an agronomist, who determined and labeled each crop type by eye. They then trained a convolutional neural network to automatically generate crop labels for the other 79,000 images, using various training methods, including iNaturalist a web-based crowdsourced biodiversity database, and GPT-4V, a multimodal large language model that enables a user to input an image and ask the model to identify what the image is depicting. For each of the 81,000 images, the model generated a label of one of four crops that the image was likely depicting rice, maize, sugarcane, or cassava.
The researchers then paired each labeled image with the corresponding satellite data taken of the same location throughout a single growing season. These satellite data include measurements across multiple wavelengths, such as a locations greenness and its reflectivity (which can be a sign of water).
Each type of crop has a certain signature across these different bands, which changes throughout a growing season, Laguarta Soler notes.
The team trained a second model to make associations between a locations satellite data and its corresponding crop label. They then used this model to process satellite data taken of the rest of the country, where crop labels were not generated or available. From the associations that the model learned, it then assigned crop labels across Thailand, generating a country-wide map of crop types, at a resolution of 10 square meters.
This first-of-its-kind crop map included locations corresponding to the 2,000 GSV images that the researchers originally set aside, that were labeled by arborists. These human-labeled images were used to validate the maps labels, and when the team looked to see whether the maps labels matched the expert, gold standard labels, it did so 93 percent of the time.
In the U.S., were also looking at over 90 percent accuracy, whereas with previous work in India, weve only seen 75 percent because ground labels are limited, Wang says. Now we can create these labels in a cheap and automated way.
The researchers are moving to map crops across India, where roadside images via Google Street View and other services have recently become available.
There are over 150 million smallholder farmers in India, Wang says. India is covered in agriculture, almost wall-to-wall farms, but very small farms, and historically its been very difficult to create maps of India because there are very sparse ground labels.
The team is working to generate crop maps in India, which could be used to inform policies having to do with assessing and bolstering yields, as global temperatures and populations rise.
What would be interesting would be to create these maps over time, Wang says. Then you could start to see trends, and we can try to relate those things to anything like changes in climate and policies.
Continue reading here:
MIT researchers remotely map crops, field by field - MIT News
Fairness in machine learning: Regulation or standards? | Brookings – Brookings Institution
Abstract
Machine Learning (ML) tools have become ubiquitous with the advent of Application Programming Interfaces (APIs) that make running formerly complex implementations easy with the click of a button in Business Intelligence (BI) software or one-line implementations in popular programming languages such as Python or R. However, machine learning can cause substantial socioeconomic harm through failures in fairness. We pose the question of whether fairness in machine learning should be regulated by the government, as in the case of the European Unions (EU) initiative to legislate liability for harmful Artificial Intelligence (AI) and the New York City AI Bias law, or if an industry standard should arise, similar to the International Standards Organization (ISO) quality-management manufacturing standard ISO 9001 or the joint effort between ISO and the International Electrotechnical Commission (IEC) standard ISO/IEC 27032 standard for cybersecurity in organizations, or both. We suggest that regulators can help with establishing a baseline of mandatory security requirements, and standards-setting bodies in industry can help with promoting best practices and the latest developments in regulation and within the field.
The ease of incorporating new machine learning (ML) tools into products has resulted in their use in a wide variety of applications, including medical diagnostics, benefit fraud detection, and hiring. Common metrics in optimizing algorithm performance, such as algorithm Accuracy (the ratio of correctly predicted to total number of attempted predicted), do not paint the complete picture regarding False Positives (algorithm incorrectly predicted positive) and False Negatives (algorithm incorrectly predicted negative), nor do they quantify the individual impact of being mislabeled. The literature has, in recent years, created the subfield of Machine Learning Fairness, attempting to define statistical criteria for group fairness such as Demographic Parity or Equalized Opportunity, which are explained in section II, and over twenty others, described in comprehensive review articles like the one by Mehrabi et al (2021).1 As the field of ML fairness continues to evolve, there is currently no one standard agreed upon in the literature for how to determine whether an algorithm is fair, especially when multiple protected attributes are considered.2 The literature on which we draw includes computer science literature, standards and governance, and business ethics.
Fairness criteria are statistical in nature and simple to run for single protected attributesindividual characteristics that cannot be the basis of algorithm decisions (e.g., race, national origin, and age, among other individual characteristics). Protected attributes in the United States are defined in U.S. federal law and began with Title VI of the Civil Rights Act of 1964. However, in cases of multiple protected attributes it is possible that no criterion is satisfied. Furthermore, oftentimes a human decision maker needs to audit the system for compliance with the fairness criteria with which it originally complied at design,3 given that a machine learning-based system often adapts through a growing training set as it interacts with more users. Moreover, no current federal law nor industry standard mandates regular auditing of such systems.
However, precedents exist in other industries for both laws and standards where risk to users exists. For example, in both the U.S. (the Federal Information Security Modernization Act [FISMA], the California Consumer Privacy Act [CCPA]) and EU (the General Data Protection Regulation [GDPR]) laws protect user data. In addition, the industry has moved to reward those who find cybersecurity bugs and report them to companies confidentially, for example through bug reward programs like Googles Vulnerability Reward Program. In this article, we propose that a joint effort between lawmakers and industry may be the best strategy to improve the fairness of machine learning systems and maintain existing systems so that they adhere to fairness standards, with higher penalties for systems that pose greater risks to users.
Before elaborating further on existing regulations, we will briefly summarize what ML fairness is and illustrate why it is a complex problem.
ML fairnessand AI fairness more broadlyis a complex and multidimensional concept, and there are several definitions and metrics used to measure and evaluate fairness in ML-based systems. Some of the most common definitions include:456
The fairness criteria should equally apply to procedural and distributive aspects.9 Other approaches are also possible; among others, the use of balanced accuracy (and its related measures)10 should be explored.
These definitions of fairness often require trade-offs, as optimizing for one may negatively impact another.11 Chouldechova (2017)12 showed that it is not possible for three group-fairness criteria to be satisfied at once, so determining the appropriate fairness metric for a specific AI system depends on the context, the domain, and the societal values at stake. This is a decision that human designers of an ML system need to make, ideally prior to the systems release to actual users. Involving several stakeholders, including users of the ML system, in deciding the fairness metric is helpful to ensure the end-product aligns with the systems ethical principles and goals.13 There is an extensive literature in ethics and psychology revolving around principles of procedural fairness.141516 Throughout the literature, procedural fairness has been broadly understood as perceived fairness of the methods used to make the decisions,17 and involves high-level principles such as correctability (if a decision is perceived as incorrect, the affected party has a mechanism to challenge it), representativeness, and accuracy (in the field of ML algorithms, this means the algorithms rely on valid, high-quality information18), among others (for a detailed explanation of procedural fairness applied to ML see Footnote 17). These principles are found in regulations; for example, the correctability principle is found in GDPR privacy law as a right to rectification and in the right to object found in both CCPA and GDPR.19
Given the rapid advances in machine learning, we recognize that legal frameworks by nations or local governments may be more difficult to develop and update. Thus, we propose first looking at industry standards, which can be used to incentivize companies to perform better while also collaborating on developing standards. Certification of a product to a standard can provide customers with a signal of quality and thus differentiate among ML-based solutions in the market.
Companies who are early adopters of a novel standard of ML fairness may be able to use that standard to gain market share as well as establish a competitive advantage compared to newcomers. For example, a company that invests early on in an auditing team for its ML system may produce more transparent software, which could be apparent to the discerning customer and thus appease concerns customers might have regarding the use of their data. Industry collaboration organizations such as Institute of Electrical and Electronics Engineers [IEEE] have developed standards for recent technologies with the help of leaders in the industry that have resulted in benefits for customers. For instance, the IEEE 802 wireless standards provided a foundation for the now-widespread Wi-Fi technology and, in the early days, provided a badge to signal to customers purchasing computers that the manufacturer complied with the latest Wi-Fi standard. Current updates to the standard enable a new wave of innovation including in areas of indoor mapping. The same incentives for standardization advocated by organizations like IEEE and ISO in manufacturing quality management, electrical safety, and communications may apply to ML.20
In the following section, we include some additional background information on standards and parallels to the cybersecurity standards that could serve as a reference for developing standards for ML fairness.
Standards are guidelines or best practices developed by industry and federal regulators in collaboration to improve the quality of products and services. Although they are often voluntary, organizations choose to adopt them to signal their commitment to security and quality, among other reasons. Some of the most prolific and widely adopted standards in the technology sector are the International Organization for Standardization (ISO) / International Electrotechnical Commission (IEC) 27000 series, the National Institute of Science and Technology (NIST) Cybersecurity Framework, and the Center for Internet Security (CIS) Critical Security Controls.
While standards are recommended, regulations are enforced. Procedures and standards must be regulatorily compliant.2122 Standards are often included by reference in U.S. federal regulations,23 which means that publishing of regulations in the Federal Register and the Code of Federal Regulations (CFR) by referring to materials already published elsewhere is lawful, as long as those materials, like international standards documents, can be accessed by the public. Since some of the standards of organizations like ISO can be hundreds of pages long and are accessible in electronic form, this approach is sensible. Regulations are legally binding rules imposed by governments or regulatory authorities that are mandatory and specify penalties, fines, or other consequences for non-compliant entities. Examples of cybersecurity and privacy regulations include the GDPR in the European Union, and the CCPA and FISMA in the United States.
The following are criteria for evaluating the use of standards and regulations:
Considering these factors, a combination of standards and regulations may be the most timely and effective approach in many cases. Regulations can establish a baseline of mandatory ML security and ethics requirements, while standards can provide guidance on best practices and help organizations stay current with the latest developments in the field.
In the United States, the process of developing cybersecurity standards often involves collaboration among the federal government, industry stakeholders, and other experts. This collaborative approach helps ensure that the resulting standards are practical, effective, and widely accepted.
One key example of this is the development of the NIST Cybersecurity Framework. NIST, an agency within the U.S. Department of Commerce, plays a key function in developing cybersecurity standards and guidelines, yet is not endowed with regulatory enforcement capabilities. NIST often solicits input from industry stakeholders, academia, and other experts to ensure that its guidance is comprehensive and current. Following the issuance of Executive Order 13636, Improving Critical Infrastructure Cybersecurity, in 2013, NIST was tasked with developing a framework to help organizations better understand, manage, and communicate cybersecurity risks.24 To do so, NIST engaged in an open, collaborative process that included the following:
These examples demonstrate the value of public-private partnerships in the development of cybersecurity standards. By involving industry stakeholders in the process, the federal government can help ensure that the resulting standards are practical, widely accepted, and effective in addressing the challenges posed by cybersecurity threats.
Regulations have also played a significant role in governing cybersecurity and privacy and have often been used to set mandatory requirements for organizations to protect sensitive information and ensure privacy.
General Data Protection Regulation (GDPR) European Union: GDPR, implemented in 2018, is widely recognized as one of the most comprehensive data protection regulations worldwide. It requires organizations to protect the personal data of EU citizens, ensuring privacy and security. GDPR has been effective in raising awareness about data protection and pushing organizations to improve their cybersecurity posture and has been seen as setting a global standard25 for user privacy. A research report on GDPR impacts published by the UK government found that GDPR compliance had resulted in more investment in cybersecurity by a majority of surveyed European businesses and that organizations generally prioritized cybersecurity following the new regulations. However, challenges include the complexity of the regulation as well as high compliance costs (further details can be found here).
Health Insurance Portability and Accountability Act (HIPAA) United States: HIPAA, passed in 1996, sets standards for protecting sensitive patient data and requires health care organizations to implement cybersecurity measures for electronic protected health information (ePHI) . Although HIPAA has been successful in improving the safeguarding of patient data, it has faced criticism for being overly complex for patients who may assume it applies to contexts where it may not offer protections (such as in health mobile apps) and for the fact that patients and caregivers of patients can have difficulty accessing necessary records.26 Furthermore, when cybersecurity breaches of private health data occur, it may be difficulty for consumers to know what options they have for recourse. The law as was written in 1996 may require updating in the face of rapidly evolving cybersecurity threats.
California Consumer Privacy Act (CCPA) United States: Implemented in 2020, the CCPA grants California consumers specific data privacy rights, such as the right to know what information is stored, as well as an option for a consumer to opt-out of data sharing. The CCPA has been praised for empowering consumers and raising the bar for privacy protection in the United States. For companies that have customers in California and other states, the CCPA has resulted in a standard for consumer privacy rights that will likely be applied by companies to other states and create dynamics that contribute to shaping the U.S. privacy regulatory framework.27 However, the CCPA does face criticism for several issues including its complexity: the nine exceptions to the right of consumers to delete data may not give consumers the protection that they expect; the burden it places on businesses; and the potential conflicts with other state or federal privacy regulations.28 Some lessons may be drawn from the current complexity of privacy laws to regulation of algorithmic fairness: Consumers may not have the time to read every opt-out notice or legal disclaimer and understand on the spot what rights they may be giving up or gaining from accepting terms of service.
Federal Information Security Management Act (FISMA) United States: Signed into law in 2002, FISMA requires each federal agency to develop, document, and implement an agency-wide program to provide information security for the information and systems that support the operations and assets of the agency. Although FISMA has led to improved cybersecurity within federal agencies, it has been criticized for being overly focused on compliance rather than continuous risk management and for not keeping pace with the evolving threat landscape. FISMA does establish generally applicable principles of cybersecurity in government, such as requiring the NIST to establish federal information processing standards that require agencies to categorize their information and information systems according to the impact or magnitude of harm that could result if they are compromised, which are codified in NIST standards. Broad principles like these are sensible for businesses to adopt as well (for example, at the business-unit or organization-unit levels).
Although regulations can be effective in setting mandatory requirements for cybersecurity and raising awareness, they may also face challenges such as complexity, high compliance costs, outdated provisions, and potential conflicts with other regulations. To address these issues, policymakers should consider periodically reviewing and updating regulations to ensure they remain relevant and effective in the face of rapidly evolving cybersecurity threats. Additionally, harmonizing regulations across jurisdictions and providing clear guidance to organizations can help alleviate some of the challenges associated with compliance. Existing cybersecurity regulations may be a template for ML fairness regulations, as well.
A similar approach can be applied to ML ethics, where regulations can set a legal framework and minimum requirements for ethical ML development and deployment, while standards can provide detailed guidance on best practices, allowing for flexibility and adaptation to new technological advancements.
Until recently, there have been no comprehensive AI-specific regulations in the United States. However, there have been efforts to establish guidelines, principles, and standards for AI development and deployment, both by the U.S. government and by various organizations and industry groups. Some examples include the following:
Executive Order on Maintaining American Leadership in Artificial Intelligence (2019): Issued by the White House in 2019, this order aimed to promote sustained investment in R&D and enhance the United States global leadership in AI. Although it did not establish specific regulations, it directed federal agencies to create a national AI strategy and develop guidance for AI development and deployment.
Defense Innovation Board (DIB) AI Principles (2019): The DIB is an advisory committee to the U.S. Department of Defense (DoD). In 2019, it released a set of ethical principles for AI in defense, covering areas such as responsibility, traceability, and reliability. Although they are not legally binding, these principles provide a basis for the ethical deployment of AI within the DoD.
Organization for Economic Co-operation and Development (OECD) AI Principles (2019): In 2019, the OECD established a set of standards for use of AI that are respectful of human rights and democratic values.
NIST AI Risk Management Framework (2023): NIST has been active in AI research and standardization efforts, focusing on topics such as trustworthy AI, AI testing and evaluation, and AI risk management. In 2023, NIST published the AI Risk Management Framework (AI RMF 1.0), which aims to provide a systematic approach to managing risks associated with AI systems. This framework, once finalized, could serve as a foundation for future AI standards and guidelines in the United States.
Partnership on AI (PAI) Tenets: The PAI is a multi-stakeholder organization that brings together industry, academia, and civil society to develop best practices for AI technologies. The partnership has published a set of tenets to guide AI research and development, including principles such as ensuring AI benefits all, prioritizing long-term safety, and promoting a culture of cooperation.
Industry-specific guidelines and standards: Several organizations and industry groups have developed their own guidelines and principles for AI development and deployment within specific sectors, such as health care, finance, and transportation. For example, PrivacyCon is an annual conference to bring together industry, government, and academia that serves the development of such guidelines in 2020 the theme of the event was health data. In a recent article, Accentures global health industry lead provided some best practices for generative AI in healthcare which are likely just a beginning of guidelines as generative AI grows in adoption in that industry. Bain and Company put together design principles for the use of generative AI in financial services, again likely a topic of growing interest in the coming years. In the transportation industry, a wide consortium of auto manufacturers, chipmakers, and other industry members put together guidelines for automated driving back in 2019. These examples of guidelines, although they are not legally binding, can help set expectations and establish best practices for AI governance within their respective industries. The National AI Institute at the Department of Veterans Affairs has also been building on and harmonizing these frameworks for trustworthy AI and operationalizing them in the health care sector.
AI-related concerns, such as data privacy and algorithmic fairness, may be addressed by existing regulations and guidelines that are not AI-specific, such as GDPR, CCPA, and guidelines on algorithmic transparency and fairness from the Federal Trade Commission (FTC). As AI continues to evolve, it is likely that more targeted regulations and standards will be developed to address AI-specific concerns and ensure ethical and responsible AI governance. Comprehensive AI auditing processes will need to be developed in a timely manner and updated periodically. Additionally, a system of incentives may be needed to encourage companies to actively develop tools to address and solve AI fairness concerns.
Standard-setting bodies work well when there are mechanisms for accountability built-in. For instance, external audit committees29 can provide an accountability mechanism as long as the audits are performed periodically (e.g., quarterly or annually) and if the auditors are not influenced by the position of those who are being audited (no revolving door scenario). To ensure accountability, such auditors may be hired by testing organizations such as the Technical Inspection Association (TUV), Underwriters Laboratories (UL), or Intertek, among others. Alternatively, auditors may be part of a volunteer community, similar to code maintainers in the Linux open-source community who control the quality of any changes to the codebase. Therefore, we suggest creating fairness audits by external auditors to the firm and codifying the type of audit and frequency in an industry standard.
We propose an approach where regulations complement industry standards by adding to them tools for enforcement, rather than as a one-size-fits-all tool. Unfortunately, firmsat least right nowdo not have a strong commercial incentive to pursue ML fairness as a goal of product development and incur additional liability in the absence of agreed upon standards. In fact, it is well known that standards may not reach their potential if they are not effectively enforced.30 Since consumers do not currently have widespread visibility into which products abide by fairness criteria and fairness criteria are not yet accessible to the general consumer since they require specialized knowledge (e.g., data science and programming skills in addition to knowing the literature), it is perhaps not feasible that a majority of consumers could themselves test for unfairness in a product or service. Furthermore, it is not yet the case that most consumers have access to training sets or company proprietary algorithms to prove whether they have been harmed by an ML system, which is required for damages under the newest regulations such as the EU AI Act (see the commentary in Heikkil, 2022). The literature on ML fairness is a complex, multidisciplinary one, so computer scientists, lawyers, ethicists, and business scholars are needed to be part of driving regulations.
Under such circumstances, it is not surprising that companies do not perceive a direct financial incentive to maintain specialized staff to audit ML fairness or supervise with a fairness-oriented goal the development of ML based products, especially in a downturn in the markets. Recently, many leading companies have unfortunately laid off a number of specialized ML fairness engineering staff, in some cases closing entire departments, which results in loss of company-specific knowledge and will mean a much slower adoption of fairness principles in industries in the future. Although regulations can provide general principles (for example, meeting at minimum an equality of opportunity fairness criterion and performing an annual fairness audit; see the guidelines by the Consumer Financial Protection Bureau (CFPB) in auditing compliance to the Equal Credit Opportunity Act [ECOA] in the United States as well as expectations regarding transparency in algorithmic decision-making when related to credit applications) and provide some consumers relief in cases of egregious violations of basic fairness criteria, they are insufficient to provide incentives to companies to perform better than such minimums and do not incentivize companies to innovate beyond meeting the regulatory requirements. Although companies may wish to implement fair ML as part of every stage of their software development processes to ensure they meet the highest ethical standards,31and we encourage that in related work,32 we recognize that people, and thus companies at large, respond to incentives and the current rules of the road are still in their infancy when it comes to ML.
Market forces can provide incentives to companies to innovate and produce better products provided the advantages of the innovation are clear to a majority of consumers. A consumer may choose a product and pay more if it satisfies a fairness standard set by a leading, recognizable standard-setting body and if the benefits of that standard are apparent. For example, a consumer may prefer a lender that markets itself as ensuring no discrimination based on subgroup fairness (i.e., combinations of categories of race, age, gender, and/or other protected attributes) if the alternatives only guarantee group-level fairness. If the consumer is aware of the higher standard this product is satisfyingfor example through a standard that the product is displayingthe consumer may choose to pay a higher price for two feature-equivalent products if one satisfies a fairness standard. Thus, we call on the industry and organizations such as the Information Systems Security Association (ISSA), the IEEE, the Association for Computing Machinery (ACM), and the ISO, among others, to invest in developing an ML fairness standard, communicate their rationale, and interact with policymakers over these standards as they deploy them into products over the next five years.
We also suggest that firms create bug bounty programs for fairness errors, in which users themselves can file bug reports with the producer of the ML system, much like what exists in cybersecurity. For example, if the ML system is a voice assistant that often misunderstands the user based on a speech impediment due to a disability, the user should be able to report that experience to the company. In another example, a user utilizing an automated job resume screening tool (as some companies now have implemented in their hiring processes) who gets consistently denied and suspects the reason may be because of a protected attribute, should be able to request a reason from the service provider. In yet another example, a mobile phone application allowing the user to test for melanoma by taking a picture should allow the user to report false positives and false negatives following consultation with a physician should such consultation prove the application misdiagnosed the user, which would allow the developers to diagnose the root cause, which may include information covered by protected attributes. A researcher or independent programmer should also be able to report bugs or potential fairness issues for any ML-based system and receive a reward if that bug report is found to reveal flaws in the algorithm or training set related to fairness.
In this report, we shared some background information on the issues of ML fairness and existing standards and regulations in software. Although ML systems are becoming increasingly ubiquitous, their complexity is often beyond what was considered by prior privacy and cybersecurity standards and laws. Therefore, we expect that what norms should specify will be an ongoing conversation requiring both industry collaboration through standardization and new regulations. We recommend a complementary approach to fairness by creating a fairness standard via a leading industry standard-setting body to include audits and mechanisms for bug reporting by users and a regulation-based approach wherein generalizable tests such as group-fairness criteria are encoded and enforced by national regulations.
Link:
Fairness in machine learning: Regulation or standards? | Brookings - Brookings Institution
5 Ways to Use AI You May Have Never Even Considered – InformationWeek
It's widely believed that AI's potential has only reached the Commodore 64 stage. In other words, the best is yet to come.
As the technology gains momentum, innovation is flourishing, with new applications seemingly limited to only its users' imagination. Consider the following five examples that show how AI will continue to surprise and transform both personal and business activities.
AI can augment and accelerate the way individuals acquire knowledge and skills by fine-tuning educational experiences to the specific needs, learning styles, and multiple intelligences of each learner, says Paul McDonagh-Smith, senior lecturer of IT at MIT's Sloan School of Management, via an email interview. "AI can employ advanced algorithms to customize educational content and feedback based on a student's unique profile and progress, leading to better educational outcomes," he explains. "This innovative application of AI has the potential to revolutionize education by making it more tailored, engaging, and accessible to learners of all backgrounds and abilities."
AI systems like GPT-3 can generate novel concepts and suggestions by analyzing large amounts of text data. "This can help spark new ideas for products, services, and business models that humans may not have thought of on their own," says Scott Lard, general manager and partner at IS&T, an information systems technology search and contingency staffing firm, via email.
Related:Why Your Business Should Consider Using Intelligent Applications
What makes this approach useful is that AI systems can consider far more possibilities and variations than a single human mind, Lard explains. "By analyzing thousands of existing ideas, it can provide fresh perspectives and out-of-the-box thinking that helps organizations innovate."
Lard suggests the best way to get started with AI-enabled idea generation is to simply ask an AI model open-ended questions about potential new ideas and concepts within a specific industry or focus area. Give the AI system as much relevant context as possible to narrow the results, he advises. Then review the generated ideas to see which offer the most potential to explore further. "You can then iterate the process by refining your questions and context to produce even better results over time."
By providing a continuous and non-judgmental presence, AI can help address the escalating demand for mental health support, says Siraj M A, director of data and analytics at project engineering firm Experion Technologies, in an email interview.
Related:AI Investments We Shouldn't Overlook
Virtual companions, powered by AI, can deliver personalized interventions tailored to individual needs and preferences, M A explains. "These AI entities go beyond mere assistance; they can collect and analyze data over time, unraveling patterns crucial for a deeper comprehension of ... mental health."
When used appropriately, AI-driven virtual therapists could surmount geographical constraints, democratizing mental health care globally, M A says. "Such solutions could ensure timely support, especially for those facing barriers due to location or a limited mental health infrastructure."
M A recommends that new adopters should start by bringing together a team of mental health and AI experts to review potential opportunities. "The team should focus on identifying the right use cases in their industry, and then identify solutions that could help with early intervention, therapy support, or diagnostic assistance," he advises. "These objectives should then be evaluated alongside the data, technology and infrastructure available to come up with a list of prioritized use cases that can be pursued."
With the assistance of AI-powered team recommendation engines we can help our hiring managers pinpoint the best candidates for a specific job, says Juan Nassiff, technical manager and solutions architect at custom software development firm BairesDev, via email.
Related:The IT Jobs AI Could Replace and the Ones It Could Create
BairesDev gets more than a million job applications every year, which is virtually impossible to sort through manually, Nassiff says via email. "We leverage complex machine-learning algorithms to match talent in our database with unique project requirements that we have a need for," he states. "Our method goes beyond the traditional use of AI in customer service or data analysis, focusing on optimizing team assembly for software development projects."
Nassiff says the approach is "incredibly useful," since it ensures an equitable and skill-focused hiring process, eliminating the biases that can occur in traditional recruitment practices. "By focusing solely on skills, professional experience per skill, and project requirements, our Team Recommendation Engine enables the assembly of highly effective teams that are tailored to specific client needs," he explains. "This not only improves project outcomes, but also significantly reduces the time and resources typically spent on recruiting and team formation."
By analyzing multiple factors, such as weather and geography, AI can help exterminators build and optimize pest control measures. "This approach is particularly useful because it allows for proactive pest management, reducing the reliance on reactive and potentially harmful chemical interventions," says Rubens Tavares Basso, CTO at pest control software provider Field Routes, via email.
Basso advises potential adopters to consider data privacy and security concerns before implementing AI pest control technology. "Additionally, businesses should be mindful of potential biases in the AI algorithms and regularly update their system to adapt to changing environmental conditions," he says. "AI in pest control software provides a forward-thinking, eco-friendly solution to managing pest issues, promoting sustainable and effective practices in agriculture and other industries."
Read the rest here:
5 Ways to Use AI You May Have Never Even Considered - InformationWeek
Groundbreaking Study Questions AUC Metric in Link Prediction Performance – Medriva
In a groundbreaking study led by UC Santa Cruzs Professor of Computer Science and Engineering, C. Sesh Seshadhri, and co-author Nicolas Menand, the effectiveness of the widely used AUC metric in measuring link prediction performance is being questioned. The researchers propose a new metric, VCMPR, which they claim offers a more accurate measure of performance in machine learning (ML) algorithms.
The Area Under the Curve (AUC) metric has been a standard tool for evaluating the performance of machine learning algorithms in link prediction tasks. However, the new research suggests that AUC fails to address the fundamental mathematical limitations of low-dimensional embeddings for link predictions. This inadequacy leads to inaccurate performance measurements, thereby affecting the reliability of decisions made based on these measurements.
The study introduces a novel metric known as VCMPR, which promises to better capture the limitations of machine learning algorithms. Upon testing leading ML algorithms using VCMPR, the researchers found that these methods performed significantly worse than what is generally indicated in popular literature. This revelation has serious implications for the credibility of decision-making in ML, as it suggests that a flawed system used to measure performance could lead to incorrect decisions about which algorithms to use in practical applications.
The findings of this research have considerable consequences for the field of machine learning. The introduction of VCMPR throws a spanner in the works, challenging the status quo and pushing ML researchers to rethink their performance measurement practices. By highlighting the shortcomings of the AUC metric, the study underscores the importance of accurate and comprehensive performance measurement tools for making trustworthy decisions in machine learning.
While the research is undoubtedly groundbreaking, its recommendations are yet to be universally accepted. The machine learning community is currently grappling with the implications of this study, with some experts supporting the switch to VCMPR, while others are apprehensive about abandoning the traditional AUC metric. However, the conversation sparked by this research is crucial, as it pushes the field towards more accurate and reliable performance measurement practices.
This research by UC Santa Cruz signifies a potential paradigm shift in the field of machine learning. By challenging the effectiveness of the AUC metric and proposing a more accurate alternative, it highlights the need for constant innovation and scrutiny in the pursuit of more reliable and trustworthy machine learning practices. Whether or not VCMPR will replace AUC as the standard performance measurement tool is yet to be seen. However, one thing is certain: this research opens up a new chapter in the ongoing endeavor to enhance the accuracy, reliability, and practicality of machine learning applications.
Read the rest here:
Groundbreaking Study Questions AUC Metric in Link Prediction Performance - Medriva
Leveraging Artificial Intelligence to Mitigate Ransomware Attacks – AiThority
Swift Strategies to Combat Ransomware Attacks and Emerge Triumphant
The famous MGM hack in Las Vegas is a prime example; the perpetrators got administrative passwords over the phone. Cybercriminals were able to take advantage of recent MOVEit vulnerabilities, which affected government agencies such as the Pentagon and the DOJ. Finding, evaluating, and deciding upon the quickest route to recovery has always been difficult. Artificial intelligence has the potential to greatly impact this area.
Organizations can quickly get back to normal operations with the help of AI, which can help understand the patterns of data corruption caused by attacks. Recognizing which files require restoration is the first step in a successful recovery. Which files have been corrupted? Which servers experienced trouble? Is it possible that important datasets have been altered? In what ways did the malware alter the files? Where can I find clean files in the backups? In the aftermath of an attack, answering these concerns while trying to restore from backups will necessitate an enormous and laborious undertaking.
Read10 AI In Manufacturing Trends To Look Out For In 2024
Organizations hit by ransomware attacks must prioritize reducing the damage it causes. The daily impact on the bottom line of organizations like hospitals, government agencies, and manufacturers when their systems are down because of ransomware is enormous. Examples abound, such as the recent assaults on MGM and Clorox, which resulted in damages amounting to hundreds of millions of dollars.
The organizations reputation takes a hit and the recovery process takes weeks, costing a pretty penny. It is crucial for intelligent recovery to validate data integrity prior to an attack happening. To keep the content clean and secure, data validation should be an ongoing process that is integrated with existing data protection procedures. Even with highly complex and hard-to-detect ransomware variations, data validation sheds light on the criminal actions that accompany these attacks.
This model is not trustworthy. The only trustworthy methodology for cybersecurity data integrity inspection is a combination of large data sets with artificial intelligence and machine learning. The bad guys have brains and are leveraging AI to their advantage more and more. When used maliciously, AI can be a potent weapon. It can identify ransomware corruption just as effectively as it can facilitate intelligent and speedy recovery. When it comes to cyberattacks, enterprises will still have a hard time recovering without AI. The ability to reduce unavailability and data loss is a benefit they reap from AI. Fortunes and company names are on the line, and the stakes couldnt be higher.
Read:Sitecore Ordercloud Delivers Limitless Commerce Capabilities
The key is to understand the distinctions between cyber recovery and catastrophe recovery. While natural disasters like floods and fires do not alter data, hackers can damage and alter entire databases, files, or even the underlying infrastructure. Relying on older backup programs for recovery frequently results in unexpected and expensive problems. Backup images can be encrypted or corrupted, or even connected to cloud-based backups might be severed, in several attacks. Cybercriminals are experts at corrupting data and backups undetected, making recovery a daunting task. Complex ransomware assaults necessitate cutting-edge methods for evaluating data integrity.
This necessitates the continual observation of millions of data points. You can learn a lot about the evolution of file and database content from these data points because they go into great detail. This kind of forensic investigation can only be handled by advanced analytics paired with AI-based machine learning.
Read:Top 10 Benefits Of AI In The Real Estate Industry
Machine learning algorithms that have been trained to identify corrupt patterns can analyze these data points and make informed conclusions regarding the integrity of the data. Artificial intelligence (AI) automation of this inspection process allows for the study of massive data sets that would be almost impossible for humans to handle. Securely unlocking devices, allowing access to bank accounts and medical information are just a few examples of the many everyday applications that use data points and AI-based machine learning. In order to guarantee safety, it depends on collecting a lot of data points.
Security flaws could be easily introduced in the absence of sufficient data points. Machine learning will unlock your phone when you hold it up to your face since it captures a lot of visual data points and has been trained to recognize your face, not your doppelgangers. For instance, the training can take into account your current and future facial appearance, including any glasses you may wear. If this procedure did not incorporate a large amount of data points, the security would be readily compromised, allowing anyone with comparable facial features to unlock the phone with ease.
[To share your insights with us as part of editorial or sponsored content, please write tosghosh@martechseries.com]
More:
Leveraging Artificial Intelligence to Mitigate Ransomware Attacks - AiThority
Vapor IO connects with Comcast on AI-as-a-service offering – Light Reading
Comcast is one of the early edge network partners for a new AI- and 5G-as-a-service offering operated by Vapor IO that will eventually become available nationwide.
Vapor IO's new micro-cloud offering, called Zero Gap AI, is underpinned by the Nvidia MGX platform with the Nvidia GH200 Grace Hopper Superchip and Supermicro's AI-optimized servers.
In use cases focused on enterprise and smart city applications, Zero Gap AI aims to deliver private 5G and GPU-based "micro-clouds" to locations such as retail storefronts, factory floors and city intersections, the company said.
Zero Gap AI initially will be available as a pilot offering in two markets Atlanta and Chicago and will tie into Comcast's network infrastructure there. That activity builds on integration work that Vapor IO and Comcast announced last year focused on tests of low-latency edge services and applications.
"Our low-latency, high-bandwidth network and connectivity services unlock a world of applications for large and small businesses, residences and mobile customers," Comcast Chief Network Officer Elad Nafshi said in a statement. "We're continuously innovating and collaborating with partners like Vapor IO to identify new ways to leverage our network, and Zero Gap AI is a unique opportunity to expand the limits of what we can do together with edge computing services."
Plans to expand
Zero Gap AI will also expand into dozens of other US markets that have access to Vapor IO's "Kinetic Grid" infrastructure, including Dallas, Las Vegas and Seattle.
With the potential to connect more cable headends to the Vapor IO fabric, Vapor IO believes its cable ambitions will extend well beyond Comcast.
"We think this is going to be a big benefit to the cable operators," Vapor IO CEO and founder Cole Crawford said, noting that the company also has "enjoyed a good relationship with CableLabs for several years."
Vapor IO's launch of Zero Gap AI follows the company's buildout of a footprint of network backbone and individual points of presence across 36 US markets (via Vapor IO's own facilities or those of its colocation partners) to put that capability alongside the radio access network (RAN).
Bringing AI to the edge
The broader idea is to bring the kind of as-a-service capability that companies expect from a cloud company without the complexities of having to figure out how to build a wireless network and build out the elements for AI and machine learning. And the inferencing capabilities of AI can then be deployed at the edge instead of on-premises to support enterprises that are operating in multiple markets or wide-scale smart city deployments. Once the AI model is trained and set up, the AI inferencing component is used to make predictions and solve tasks.
"Generative AI costs you money. Inferencing makes you money," Crawford said. "And the inferencing action, I think, is where the industry will make a lot of money."
Vapor IO didn't announce any early, specific deployments of Zero Gap AI, but the company did spell out several potential use cases. A retailer, for instance, could use it for AI-assisted automated checkouts without having to deploy expensive AI gear at each store.
'Computer vision' a driving force
In another example, a city could use Zero Gap AI for "computer vision" services to support a pedestrian safety system across hundreds of intersections without having to deploy AI equipment at every corner. Additionally, construction sites could use computer vision with AI inferencing to determine if everyone working there is wearing a hardhat.
In a more specific example, the City of Las Vegas is working on computer vision inferencing capabilities that would take advantage of thousands of cameras deployed around the city for use in areas such as public safety, law enforcement and traffic management.
"All applications we are seeing demand for today use computer vision in some form or another. Computer vision certainly is the biggest driver," Matt Trifiro, Vapor IO's chief marketing officer, said.
"We think large vision models as a basis for how to do inferencing are going to generate more revenue for the industry than large language models, I think, over the next five years," added Crawford.
Read the original:
Vapor IO connects with Comcast on AI-as-a-service offering - Light Reading
The world’s coral reefs are even bigger than we previously thought – BGR
Our worlds coral reefs are much larger than we previously believed. According to a report shared on The Conversation by Mitchell Lyons, a postdoctoral research fellow at The University of Queensland, and Stuart Phinn, a Professor of Geography at The University of Queensland, researchers found 64,000 square kilometers of coral reef we didnt know existed.
The ground-breaking discovery brings the total size of our planets shallow reefs to roughly 348,000 square kilometers. Thats roughly the size of Germany, the two researchers note in their report. This new figure fully represents the worlds coral reef ecosystems, including coral rubble flats, as well as living walls of coral.
Whats even more astounding about this discovery is that it was mostly only made possible thanks to machine learning. The researchers say that they relied on snorkels, satellites, and machine learning to help them discover the hidden coral reefs. These high-resolution satellites made it possible to view reefs as deep as 30 meters down.
When coupled with the direct observations and records of the worlds coral reefs, the researchers say they were able to ascertain that a large amount of the worlds coral reefs had not been identified or noted down anywhere. And that had to be changed.
Sign up for the most interesting tech & entertainment news out there.
By signing up, I agree to the Terms of Use and have reviewed the Privacy Notice.
They used machine learning techniques to help create new maps of the coral reefs found around the world, while they relied on satellite imagery and data to create predictions that were as accurate as possible when producing the new maps. Of course, without direct observational data, its hard to confirm the existence of these reefs fully.
But, it is still a huge step forward for the study of the worlds coral reefs, which are constantly in danger due to the ongoing climate change issues plaguing our planet. Perhaps with more useful studies like this, we can get a better understanding of how much is truly at stake.
See the rest here:
The world's coral reefs are even bigger than we previously thought - BGR
Machine learning driven methodology for enhanced nylon microplastic detection and characterization | Scientific Reports – Nature.com
Contamination level and representativeness of subsampled areas
Analysis of the procedural blank sample indicated that contamination from the experiment environment was low. The detailed results are summarized in Table S1.
The positive control sample was used to examine the representativeness of the nine subsampled regions, determined by the ratio of the estimated mass of particles on the filter to the initial 0.05mg of nylon microspheres. It was found that the method slightly overestimated the mass of nylon microspheres, with an obtained ratio of 1.270.06. (Detailed information is displayed in Fig. S1). In an endeavour to optimize the number of regions for our spectral imaging model, we systematically analyzed the ratio between estimated and actual values across varying numbers of regions.
Generally, for identification of MP using the O-PTIR technique, there are three commonly used methods, i.e., DFIR imaging, point spectra measurements, and HSI. However, each of these methods brings some challenges: for example, DFIR imaging is fast yet provides unreliable results while HSI and point spectra measurements allow for accurate results, but they are time-consuming for data collection. With the QCL system integrated within the O-PTIR microscope, the microscope can generate a single frequency IR image of a 480m640m (spatial resolution: 2m) area of a filter in approximately 3min and 20s. When an appropriate wavenumber and a threshold value are selected, the generated image shows the majority of MP particles while ruling out most non-MP particles. With this method however, careful selection of a suitable wavenumber and a threshold value for MP particles are necessary; multiple threshold values might be needed in case of interference from the complex non-MP particles. In our study, the discrimination between MPs and non-MP particles based on single-wavenumber images proved to be unfeasible, as illustrated in Fig. S2.
The second method commonly used for MP identification is point spectra measurements. After particles are observed in the mIRage microscope, point spectra could be collected for each particle and compared against parent plastic to achieve chemical identification of the particles. This method presented two challenges: (1) When using visible light for particle location under the microscope, non-MP particles were inevitably included as spectral acquisition targets, thus adding to the analysis time. (2) For an individual particle, the O-PTIR spectra could vary significantly across different spots of the particle (see example in Fig.1). This necessitates the collection of spectra from multiple areas of the particle to enhance the reliability of identification results. Consequently, the analysis time will be multiplied. For example, it takes 25s to obtain a spectrum (a total of 5 scans acquired for each single spectrum), so if there are 100 particles from the regions of interest on the filter and three spectra are required for each particle, the total analysis time needed is at least 2h. This estimation only accounts for the raw data acquisition, excluding additional durations associated with manual adjustments such as repositioning the objective or refocusing. In light of this, such an approach becomes exceedingly time-intensive, especially when a vast number of particles are in play.
Two spots of the particle encircled in a red dashed line (A) selected for point spectra collection and (B) the corresponding O-PTIR spectra of the two spots. a.u. is arbitrary units. The scale bar is 20m.
HSI was the third method employed for MP identification. HSI generates an image where each pixel contains a full spectrum. Hence, it is a reliable method for MP identification. However, this reliability comes at the cost of drastically longer data collection time, which makes HSI impractical for routine MP analysis. For example, capturing a hyperspectral image for a 480m640m area (spatial resolution of 2m and spectral resolution of 2cm1, from a spectral range of 7691801cm1) requires almost two weeks.
In response to the challenges mentioned above, we have developed a reliable MP detection framework with an improved speed that is suitable for detecting a large quantity of nylon MPs. It can collect spectral data from nine areas (the size of each area is480m640m) of a filter (at a spatial resolution of 2m) within just approximately 2h. Powered by machine learning, the reliability of this framework is not compromised in response to reduced data collection time.
In order to effectively utilize DFIR imaging for high-throughput analysis of MPs, it is crucial to carefully select specific wavenumbers that provide the greatest discriminatory power between MP and non-MP particles. Making incorrect choices in wavenumber selection can directly impact the accuracy of identification. Acquiring too many wavenumbers increases measurement time, resulting in decreased throughput. For instance, adding just one more wavenumber can lead to an approximate 30-min increase in the time required for our proposed MP detection framework to collect data from a single filter. To identify the important wavenumbers and determine the optimal number of such wavenumbers, a database collected from bulk nylon plastic was assembled, containing 1038 spectra of MP and 1052 spectra of non-MP.
We found several types of non-MP particles in our dataset. Figure2 displays the spectra of two non-MP classes (type I non-MP and type II non-MP), along with the mean spectrum of MP, enabling a comparison. Upon initial inspection, type I non-MP exhibits a prominent sharp peak in the 17001800cm1 spectral range, while type II non-MP displays a broad peak in the 10001200cm1 spectral range. In contrast, the apparent characteristic peaks of MPs are two consecutive sharp peaks in the 15001650cm1 range.
Mean spectra for nylon MP class and two non-MP types from the database constructed, following standard normal variate (SNV) to minimize the multiplicative effects.
Two thirds of the spectra from each class were randomly selected as the training dataset for model development, and the remaining samples formed the test dataset. Based on the obtained results, the model utilizing the full wavenumber spectrum yields a correction accuracy rate of 85.31% (see Table 2). The confusion matrix of the SVM-Full wavenumber model (Fig.3A) implies that there are 8 point spectra of MPs wrongly classified as non-MPs and 97 of non-MPs mistakenly assigned as MP.
Confusion matrix showing classification accuracy for the test set of SVM-Full model using full spectral variables (A) and SVM-Four model (B).
Subsequently, the coefficient based feature importance for the full wavenumber model (Fig.4) was plotted to visualize the contribution of individual spectral variables. According to Fig.4, we could choose the important wavenumbers to our dataset based on the feature importance. The higher feature importance signifies stronger discriminative capability. Based on the analysis of the coefficients of the SVM-Full wavenumber model, wavenumbers 1711cm1, 1635cm1, 1541cm1, and 1077cm1 (indicated in Fig.4) showed the feature importance, hence, were selected as important wavenumbers for distinguishing between MPs and non-MPs. As seen from Table 2, the model optimized with these four wavenumbers demonstrates an enhanced correction rate of 91.33%. Meanwhile, the SVM-Four wavenumbers model (Fig.3B) resulted in 34 point spectra of MPs wrongly classified as non-MPs and 28 of non-MPs mistakenly assigned as MP, which shows it is a balanced model for classification tasks. The SVM-Four wavenumbers model appears to outperform the SVM-Full wavenumber model in terms of specificity, CCR, and MCC, suggesting that it is a better model for this classification task. However, the SVM-Full wavenumber model has a higher sensitivity, making it better at identifying true positive cases.
The coefficients (or weights) of the SVM model, which indicate the importance of each feature (wavelength), are then plotted. Four wavenumbers which has relatively higher feature importance than other are marked above the curves (i.e., 1711cm1, 1635cm1, 1541cm1, and 1077cm1).
After the selection of the four important wavenumbers, DFIR images were obtained at the important wavenumbers from the nine subsampled regions of the filter. Particle identification could be performed through visual inspection of these DFIR images. For instance, Fig.5A shows an optical image of a small region of a filter with a particle in the centre, and Fig.5B shows chemical images of that region based on the intensity of 1711cm1, 1635cm1, 1541cm1, and 1077cm1 bands. The absorbance intensity of each chemical image was normalized to the same range. The particle in this region exhibits high signal intensity at 1635cm1 and 1541cm1, while showing weak signal intensity at 1711cm1 and 1077cm1, indicating that it is a MP particle. On the other hand, non-MP particles would show weak signal intensity at 1635cm1 and 1541cm1, while showing strong signal intensity at 1711cm1 and/or 1077cm1 (See Figs. 6A,B for an example of non-MP particles).
An optical image of an area of a prepared filter, with a MP particle in the center of the image (A), single frequency images of that area using 1711cm1, 1635cm1, 1541cm1 and 1077cm1 band intensity, with the absorbance intensity of each chemical image normalized to the same range (B), support vector machine (SVM) prediction results of the particles in this area (C), and normalized O-PTIR spectra of the particle and the bulk plastic (D). The +1 in (C) indicates where the spectrum of the particle in (D) was collected. The scale bar is 20m.
An optical image of an area of a prepared filter, with a non-MP particle in the center of the image (A), single frequency images of that area using 1711cm1, 1635cm1, 1541cm1 and 1077cm1 band intensity, with the absorbance intensity of each chemical image normalized to the same range (B), support vector machine (SVM) prediction results of the particles in this area (C), and normalized O-PTIR spectra of the particle and the bulk plastic. The +1 in (C) indicates where the spectrum of the particle in (D) was collected. The scale bar is 20m.
However, for accurate particle identification, visual inspection is not advisable due to low accuracy. Meanwhile, application of SVM-Full model requires a huge amount of time in the collection of point spectra from all particles. Therefore, an SVM-Four wavenumbers model was trained from the four important wavenumbers to predict each particle accurately. Spectral data at the four important wavenumbers were extracted from the same database used for the SVM-Full wavenumber model. The trained SVM model on the selected four wavenumbers demonstrated good performance, evidenced by a high CCR, MCC, sensitivity and specificity (Table 2).
After applying the SVM classifier to the particle in Fig.5A, each pixel of the particle was labelled as either MP (red) or non-MP (blue), providing an intuitive and accurate identification result. Figure5C displays the SVM prediction results for one example area. As can be seen, most pixels in the particle have been labelled as MP, with a small portion labelled as non-MP. The result for a particle was determined by the majority vote of the labels of all pixels within the particle. Thereby this particle was identified as a MP particle. This was further confirmed by the full spectrum of this particle (Fig.5D). Also, by applying the SVM classifier to the particle in Fig.6A, the particle was predicted to be a non-MP particle (Fig.6C). Figure6D presents a spectrum of this particle, which validates the predicted outcome.
Our developed SVM model offers several distinct advantages over the traditional correlation-based method for MP identification. Firstly, the SVM model only requires four wavenumbers as input, significantly reducing the complexity of data collection compared to the correlation-based approach, which involves obtaining spectra from each particle and calculating correlation coefficients. This efficiency translates into a substantial time-saving advantage. Therefore, the developed method is particularly useful when dealing with a large number of particles on the filter. Secondly, the correlation-based method often relies on establishing a threshold for identification, introducing a subjective element into the process. In contrast, the SVM model automates the assignment of particles to MP or non-MP categories, contributing to a more consistent and reliable MP identification process. Last but not least, once essential wavenumbers are identified and a simplified model is developed, the SVM approach can be extended to identify a range of polymers. This versatility is a significant advantage, enabling the model to adapt to various MP compositions beyond the scope of the original correlation-based method.
Using the novel identification procedure developed, it was possible to investigate the effectiveness of several sample pre-processing steps in a more representative and less biased and efficient way. To this end, high-temperature filtration and alcohol pretreatment were chosen as methods for reducing non-MP. The performance of these two treatments was evaluated separately, including the analysis of the spectra and DFIR images at four selected wavenumbers. The evaluation included an assessment of their impact on the spectra of MP and their effectiveness in removing non-MP. To assess the effectiveness of particle removal, the MP particle/all particle ratio (MP/All) detected by four wavenumbers SVM model was used. A treatment was considered effective if it significantly increased this ratio.
By boiling the nylon bulk, MP particles were released. The released particles were subsequently enriched on the filters through high-temperature filtration and room-temperature filtration, respectively. The mean spectrum of MP from high-temperature filtration, the mean spectrum of MP from room-temperature filtration, and the mean spectrum of nylon bulk were plotted together for comparison (Fig. S3). Results showed that when the mean spectrum of nylon bulk was compared to the mean spectra of MP (regardless of the filtration temperature), no consistent peak shift was found. When the mean spectrum of MP from high-temperature filtration and the mean spectrum of MP from room-temperature filtration were compared, no consistent peak shift was found either. These findings demonstrate that exposure to high temperatures reaching water boiling point will not impact the spectral profiles of MPs when compared to the original bulk plastic.
After the thermal degradation of nylon bulk, the particles released were captured on filters through high-temperature filtration and room-temperature filtration, respectively. Using our developed SVM classifier, particles in the nine subsampled regions of the filter were counted and subsequently the ratio MP/All was calculated. The MP/All ratio from the room-temperature filtration was 0.0900.012, and from the high-temperature filtration was 0.080.012, respectively. The normal t-test results indicated that the effectiveness of high-temperature filtration in removing non-MP was not evident.
Gerhard et al.18 reported that slip agents (such as fatty acid and fatty acid esters) of plastic products are released concomitantly with the release of MP particles, and these slip agents might be dissolved in hot water and washed away during the filtration process. In light of this, our results suggest that the nylon bulk used in our study might have just a small amount fatty acid or their esters. Indeed, Hansen et al.19 reported that as additives in plastics, the amount of slip agents could be as low as 0.1%, and the removal of a small amount of additives from MP samples might not statistically significant. Furthermore, based on observations of the prepared filters, we did not see a thin residue on the room-temperature filter, which was observed by Gerhard et al.18 who confirmed that most part of the thin residue in their experiment was identified as additives. This supports that the amounts of hot water-rinseable additives in our samples were low, however this would generally be sample specific.
After the degradation of nylon bulk in boiling water, the particles released were retained on filters. An alcohol treatment was subsequently applied to the filters to reduce non-MP particles. The mean spectra of MP before and after an alcohol treatment and the mean spectrum of nylon bulk were plotted together and compared (Fig. S4). Results revealed that when the mean spectrum of nylon bulk was compared to the mean spectra of MP (regardless of the alcohol treatment), no consistent peak shift was observed. When the mean spectra of MP before and after the alcohol treatment were compared, no consistent peak shift was observed either.
To further explore the effects of alcohol treatment on released particles, this paragraph focuses on spectral changes of individual particles. The spectral data of individual particles was baseline corrected, smoothed, and normalized to between 0 and 1 prior to comparison. Figure7 shows spectra as well as optical images of 4 MP particles before and after an alcohol rinse. For all four particles presented, peak shifts for signature bands of MP in the range of 769cm1 to 1801cm1 were not observed. Particle 1 has a peak at 1741cm1 before the alcohol treatment; this is a peak that has been assigned to the formation of carbonyl groups during polyamide 66 photo-20 and thermal-oxidation21, which implicates a pathway of oxidation in hot water for the particles during high temperature treatment (filtration at 70C). However, the reduction in signal intensity of this peak after the alcohol treatment might indicate that the alcohol treatment could remove some of the oxidized substances. The spectrum of particle 2 has two new peaks at 1007cm1 and 1029cm1, respectively, after exposure to alcohol, which was possibly due to alcohol residue, as these two new peaks correspond to C=O stretching bonds of alcohol22. No introduction or disappearance of the peak was observed in the spectra of particle 3 and particle 4. By observing the optical images of these MP particles, it can be concluded that alcohol treatment did not have an effect on their morphology.
Optical images of nylon MP particles 1, 2, 3, 4, with the particles circled and marked with numbers. The scale bar is 10m.
Figure8 shows spectra as well as optical images of 4 non-MP particles before and after an alcohol rinse. Particle 1 and particle 2 appear to be yellowish to brownish. These types of non-MPs are easy to be discriminated against from MPs based on visual observation of optical images, as most of the MP particles in our experiments are whitish, similar to the color of their bulk plastic samples. However, judgement based on color is not always correct. Subsequent spectral analyses confirmed that particle 1 and particle 2 are not MP. After the alcohol treatment, most parts of these two particles were washed away, leaving black remnants on the filter. Though the elimination was not complete, it proved that alcohol could remove non-MP particles. Particle 3 is whitish with a glossy surface, and it is a chlorinated polyethylene particle. After the alcohol treatment, particles with a spectrum similar as chlorinated polyethylene (We do not have any appliances containing polyethylene) remained where it had been, and the spectrum was not changed substantially. The glossiness of the particle was reduced; however, this indicates that alcohol treatment could not remove this type of contaminant. Particle 4 is a white particle, and it is covered by a brown, lumpy object on the upper left. The noise spectrum cannot be identified by the database with high certainty. After the alcohol treatment, it appeared dull grey, and its spectrum looked like that of nylon showing five signature peaks (1633cm1, 1533cm1, 1464cm1, 1416cm1, 1370cm1). This implies that the alcohol might be able to remove some contaminants, such as additives, which cover the surface of the MP particle. Li et al.23 have reported the same finding that alcohol could wash away some additives attached to the surface of MP particles. The above experiments prove that an alcohol treatment could remove some particle contaminants and wash away some impurities covering MP particles.
Optical images of non-MP particles 1, 2, 3, 4, with the particles circled and marked with numbers. The scale bar is 10m.
To further explore the significance of alcohol treatment, the developed SVM classifier was used to count the particles in the nine subsampled regions of the filter, based on which the MP/All was calculated. The MP/All ratio before the alcohol treatment was 0.1290.129; and after the alcohol treatment was 0.2860.207, respectively. The paired t-test of the data indicates that an alcohol treatment of the same areas of the filter significantly increases the MP/All (p<0.05). In summary, alcohol treatment was significantly effective in reducing non-MP contaminants.
The proposed MP detection framework was specifically adapted for application to detect MPs released from nylon teabags. However, it's important to note that not the entire framework was employed in this context. Rather, a selective application was implemented, excluding the components based on DFIR imaging and the SVM-Four wavenumber model. After steeping teabags in hot water, MPs were released and collected on a filter through filtration at room temperature. This filter was rinsed with alcohol and air-dried in the fume hood prior to O-PTIR data collection. The contaminants from the teabag are not the same as those originating from reference nylon bulk plastics. For example, teabags might have some contaminants from tea residuals, as noted by Xu et al.13. Particles released from nylon teabags were identified through point spectra measurements due to the relatively low particle count (i.e., <5 particles) observed in the subsampled regions of the filter (see Conventional MP identification).
Characterization of MP particles released from teabags was carried out using the MATLAB image processing toolbox function regionprops, which calculates properties of each particle including area, length (length of the major axis of the fitted ellipse), width (length of the minor axis of the fitted ellipse), and circularity. In Fig.9, we present four optical images toshow nylon MP particles, which have been released from three nylon teabags; they are circled and marked with numbers. To provide a comprehensive analytical context, the spectra of three key references are plotted alongside: a nylon reference sphere, a sample of nylon in bulk form, and the material of the nylon teabag itself. This juxtaposition allows for a direct comparison between the spectra of the isolated particles and these standard nylon references, this contributes to a more detailed understanding of the appearance as well as the spectral properties of the particles.
Optical images of nylon MP particles 1, 2, 3, 4 released from nylon teabag, with the particles circled and marked with numbers.
The average quantity of MP in the nine subsampled regions of the filter was 8.71.2. Extrapolating to the whole filter, we would estimate 31943.7 MP particles released from steeping three teabags, or approximately 106.314.6 MP particles were released from one teabag. The particle counts/quantities of MPs released from teabags previously reported are listed in Table 1. Our reported count is comparable to that reported by Ouyang et al.9, who found 393 MPs using FTIR-based particle-based analysis, although their brewing time was much longer than ours (1h vs 5min). Regarding Hernandez et al.7, their brewing temperature and time are very similar to ours. Nevertheless, as they did not conduct particle-based analysis, their results were overestimated8. The detection limits of O-PTIR spectroscopy and Raman spectroscopy are similar, with O-PTIR spectroscopy being around 500nm and Raman spectroscopy being around 1m. Based on this, we were surprised to find that the number of MPs we detected was one to two orders of magnitude lower than the 5,800 20,400 per teabag (brewed at 95C for 5min) reported by Busse et al.8 using Raman spectroscopy. Busse et al.8 conducted particle-based analysis, indicating that their results should be considered reliable. However, it is important to notethat their use of Raman spectroscopy may have led tomisidentification of non-MP particles as MPs in an unexpected way. To illustrate, Busse et al.8 identified and counted polyethylene (PE) particles in the teabag leachate. However, these PE particles could also be behenamide (CH3(CH2)20CONH2), which is a typical slip additive widely used in PE plastic. Behenamide exhibits a high level of spectral similarity with PE in Raman spectroscopy, up to 90%, mainly due to the strong Raman signal associated with its saturated alkyl chains (i.e., (CH)) and relatively weak Raman signals from carbonyl and amine groups23. The observed disparities between our results and those of Busse et al.8 could also potentially be attributed to the use of different types of teabags.The counts/quantities reported by other studies listed are expressed in the mass of MPs released per teabag11,12, or the number of MP particles per kg of teabags10. Therefore, direct comparisons with these studies are not possible in our paper.Subsequently, the length, width, area, and circularity of each particle were measured and calculated using the MATLAB function regionprops. Figure10A shows the surface area of the MP particles. Except for the two MP particles with the smallest (100m2) and largest (680m2) surface area, the majority of the remaining particles have surface areas ranging from 150 to 550m2. Figure10B shows the distribution of the length of MP particles. As can be seen, the maximum length is 40m and the minimum length is 16m, while most MP particles have a length ranging from 18 to 28m. Figure10C displays the width of the MP particles. As seen from the graph, the smallest width is 9m, while the largest width is 30m. The majority of the MPs have a width range between 12 and 24m. Figure10D shows the circularity of the MP particles. Among all the MP particles, only 4 have a low circularity (0.10.4), while most of the MP particles have circularity ranging from 0.65 to 0.95. Circularity is a measure of how closely a shape resembles a perfect circle. Circularity values near 1 represent perfect circles, while values close to 0 indicate shapes that deviate significantly from circularity. Based on the literature, particles that are more circular in shape are found to be less toxic, while those that deviate from a circular shape, manifesting more stretched or fiber-like, are associated with a higher level of toxicity24.
Length (A), width (B), area (C) and circularity (D) of MP particles released from steeping a single teabag.
View original post here:
Machine learning driven methodology for enhanced nylon microplastic detection and characterization | Scientific Reports - Nature.com
Revolutionizing Healthcare: The Impact of Machine Learning | by NIST Cloud Computing Club | Feb, 2024 – Medium
Thanks to technology breakthroughs, the healthcare business has undergone a dramatic transition in recent years. Machine Learning (ML) is at the vanguard of this revolution. Artificial intelligences subset of machine learning is revolutionizing the healthcare industry with the promise of better diagnosis, individualized treatment plans, and more effective healthcare systems.
Machine learning has made significant advances in healthcare, one of which is its unmatched speed and accuracy in analyzing large volumes of medical data. Machine learning algorithms has the ability to sort through genomic data, medical imaging, and electronic health records, revealing patterns and connections that human eyes would miss. This capacity is particularly important for early illness diagnosis and detection.
For example, in radiology, ML algorithms are enhancing the accuracy of medical imaging interpretations. They can quickly analyze complex medical images like MRIs and CT scans, aiding radiologists in detecting abnormalities and identifying potential health issues. This not only expedites the diagnostic process but also improves the precision of medical diagnoses.
The idea of individualized medicine is being revolutionized by machine learning. With the use of individual patient data analysis, including lifestyle factors, genetic information, and therapy responses, machine learning algorithms can customize therapies to meet the specific needs of every patient. This method is more focused, reducing side effects and maximizing the effectiveness of treatment.
In treatment of Cancer, for example, ML is being employed to predict how specific cancer types will respond to various treatment options based on genetic markers. This enables oncologists to recommend personalized treatment plans, improving the chances of successful outcomes and reducing the need for trial-and-error approaches.
Beyond diagnostics and treatment, Machine Learning is also playing a pivotal role in optimizing healthcare systems. Predictive analytics can forecast patient admission rates, enabling hospitals to allocate resources efficiently. ML algorithms can identify trends in patient data to anticipate disease outbreaks, allowing for proactive public health measures.
Furthermore, ML-powered chatbots and virtual health aides are revolutionizing patient relationships. These solutions promote more easily available and convenient healthcare services by offering real-time monitoring for patients with chronic diseases, scheduling appointments, and giving prompt answers to health-related questions.
Although machine learning has bright futures in healthcare, there are obstacles and moral issues to be addressed. Careful thought must be given to matters like algorithm bias, data privacy, and the interpretability of machine learning models. Ensuring the proper implementation of machine learning (ML) in healthcare requires striking a balance between innovation and ethical principles.
In conclusion, the integration of Machine Learning in healthcare is reshaping the industry, from diagnostics to personalized treatment and system optimization. As these technologies continue to advance, they hold the potential to revolutionize patient care, improve outcomes, and usher in a new era of precision medicine. While challenges persist, the ongoing collaboration between healthcare professionals, data scientists, and policymakers is essential for realizing the full benefits of Machine Learning in healthcare.