Category Archives: Machine Learning

Know Labs Demonstrates Improved Accuracy of Machine Learning Model for Non-Invasive Glucose Monitor – Marketscreener.com

SEATTLE - Know Labs, Inc. (NYSE American: KNW) today announced results from a new study titled, 'Novel data preprocessing techniques in an expanded dataset improve machine learning model accuracy for a non-invasive blood glucose monitor.'

The study demonstrates that continued algorithm refinement and more high-quality data improved the accuracy of Know Labs' proprietary Bio-RFID sensor technology, resulting in an overall Mean Absolute Relative Difference (MARD) of 11.3%.

As with all Know Labs' previous research, this study was designed to assess the ability of the Bio-RFID sensor to non-invasively and continuously quantify blood glucose, using the Dexcom G6 continuous glucose monitor (CGM) as a reference device. In this new study where data collection was completed in May of 2023, Know Labs applied novel data preprocessing techniques and trained a Light Gradient-Boosting Machine (lightGBM) model to predict blood glucose values using 3,311 observations - or reference device values - from over 330 hours of data collected from 13 healthy participants. With this method, Know Labs was able to predict blood glucose in the test set - the dataset that provides a blind evaluation of model performance with a MARD of 11.3%.

Comparatively, Know Labs released study results in May 2023 that analyzed data from five participants of a similar demographic using 1,555 observations from 130 hours of data collection, and the first application of the lightGBM ML model, which resulted in an overall MARD of 12.9%.

In June 2023, Know Labs announced the completed build of its Gen 1 prototype, which incorporates the Bio-RFID sensor that Know Labs has been using to conduct clinical research in a lab environment for the last two years, and has published results of its proven stability, into a portable device. Testing with the Gen 1 device is underway, optimizing the sensor configuration for data collection, including new environmental and human factors.

The Company's focus is on collecting more high-quality, high-resolution data across a diverse participant population representing different glycemic ranges and testing scenarios, to refine its algorithms based on this new data, and to optimize its sensor in preparation for scale. To support this work, the Company is continuing to test with its Gen 1 device every day in parallel with ongoing clinical research with its stationary lab system. Gen 1 is expected to generate tens of billions of data observations to process which will be critical to helping validate algorithm performance across the real-world scenarios in which Know Labs' glucose monitoring device may be used. This is a key component of realizing the Company's vision for bringing an FDA-cleared product to the market.

About Know Labs, Inc.

Know Labs, Inc. is a public company whose shares trade on the NYSE American Exchange under the stock symbol 'KNW.' The Company's technology uses spectroscopy to direct electromagnetic energy through a substance or material to capture a unique molecular signature. The Company refers to its technology as Bio-RFID. The Bio-RFID technology can be integrated into a variety of wearable, mobile or bench-top form factors. This patented and patent-pending technology makes it possible to effectively identify and monitor analytes that could only previously be performed by invasive and/or expensive and time-consuming lab-based tests. The first application of our Bio-RFID technology will be in a product marketed as a non-invasive glucose monitor. The device will provide the user with accessible and affordable real-time information on blood glucose levels. This product will require U.S. Food and Drug Administration clearance prior to its introduction to the market.

Safe Harbor Statement

This release contains statements that constitute forward-looking statements within the meaning of the Private Securities Litigation Reform Act of 1995 and Section 27A of the Securities Act of 1933, as amended, and Section 21E of the Securities Exchange Act of 1934, as amended. These statements appear in a number of places in this release and include all statements that are not statements of historical fact regarding the intent, belief or current expectations of Know Labs, Inc., its directors or its officers with respect to, among other things: (i) financing plans; (ii) trends affecting its financial condition or results of operations; (iii) growth strategy and operating strategy and (iv) performance of products. You can identify these statements by the use of the words 'may,' 'will,' 'could,' 'should,' 'would,' 'plans,' 'expects,' 'anticipates,' 'continue,' 'estimate,' 'project,' 'intend,' 'likely,' 'forecast,' 'probable,' 'potential,' and similar expressions and variations thereof are intended to identify forward-looking statements. Investors are cautioned that any such forward-looking statements are not guarantees of future performance and involve risks and uncertainties, many of which are beyond Know Labs, Inc.'s ability to control, and actual results may differ materially from those projected in the forward-looking statements as a result of various factors. These risks and uncertainties also include such additional risk factors as are discussed in the Company's filings with the U.S. Securities and Exchange Commission, including its Annual Report on Form 10-K for the fiscal year ended September 30, 2022, Forms 10-Q and 8-K, and in other filings we make with the Securities and Exchange Commission from time to time. These documents are available on the SEC Filings section of the Investor Relations section of our website at http://www.knowlabs.co. The Company cautions readers not to place undue reliance upon any such forward-looking statements, which speak only as of the date made. The Company undertakes no obligation to update any forward-looking statement to reflect events or circumstances after the date on which such statement is made.

Contact:

Laura Bastardi

Email: Knowlabs@matternow.com

Tel: (603) 494-6667

(C) 2023 Electronic News Publishing, source ENP Newswire

Read the original post:
Know Labs Demonstrates Improved Accuracy of Machine Learning Model for Non-Invasive Glucose Monitor - Marketscreener.com

Research hotspots of deep learning in critical care medicine | JMDH – Dove Medical Press

Introduction

Deep learning (DL) is a subset of machine learning (ML) that is created using complex algorithms that are inspired by the organization of the human brain with many discrete nodes or neurons and can identify important patterns or features in a dataset.1 DL and ML refer to two different technologies, and DL is considered an advanced structure of ML. Convolutional neural networks, long and short-term memory networks, recurrent neural networks, transformer models, and attention mechanisms are all common u DL technologies.2 ML techniques are a collection of mathematical and statistical concepts such as support vector machine, random forest, and K-nearest neighbors.3 Whereas DL algorithms are specialized techniques that are a subset of ML.1 The most important difference between the two approaches is that ML requires a feature engineering process that eliminates unnecessary variables and pre-selects only those that will be used for learning.4 This process is disadvantaged by the requirement that experienced professionals pre-select critical variables. Conversely, DL algorithms overcomes this shortfall by a process in which have built-in mechanisms for assessing and addressing the root of any inaccuracies and do not require guidance.4 To interpret an image, for example, to deconstruct the image into specific features, such as brightness, curvature, sharpness, etc., the extraction process of the support vector machine, a ML technique, requires digital input into the computer algorithms to extract these image features.5 Whereas this feature extraction process is completely different in DL algorithms. By varying the weights of the given features, DL applies a series of convolutional filters to the image, and the DL algorithm can then be trained to recognize a specific type of image and ultimately achieve the extraction of features from the image.6

Patients in the critical care medicine (CCM) field usually have aggressive and complex conditions, complicated medical record data, and clear trends in personalized treatment, resulting in a huge need for automated and reliable health information processing and analysis.7 DL algorithms and models enable machines to mimic human activities such as seeing, hearing, and thinking, helping to solve many complex pattern recognition challenges, and such features seem to help bring breakthroughs to the CCM. In addition to its well-known use in image processing and analysis, DL is also widely used in the medical field for health-record analysis, clinical diagnosis, health monitoring, personalized medicine, and drug development.7 Managing most diseases in the field of CCM, such as sepsis, acute respiratory distress syndrome, and severe stroke, largely requires the application of DL technologies.810 Currently, various algorithms and models based on DL have been widely used in the management of common diseases in CCM (such as sepsis and acute respiratory distress syndrome), including early detection of diseases, severity score estimation, facilitating ICU liberation through early successful extubation, early mobility, and survival prediction.1115 Accumulating evidence suggests that the application of DL technology promotes intelligence in CCM, not only effectively improving the quality of medical care, but also helping to increase the efficiency of clinicians.16

Publications on the application of DL in CCM have continued to grow in recent years. The continuous increase in publications is positive for the updating of knowledge, but it also poses a challenge for researchers, as the process of acquiring knowledge makes it difficult to avoid the heavy work of combing publications.17 As a quantitative research method used to analyze the scholarly characteristics of the literature in certain scientific fields, bibliometrics helps researchers to grasp research hotspots and trends in their fields of interest and to predict their prospects.18 Therefore, the present bibliometric and visualized study was performed to provide a comprehensive overview of current hotspots, future trends in the use of DL in CCM, to showcase the contributions of leading countries, authorities, and prominent scholars, and to provide clues to potential future collaborations and research directions.

Data were obtained from the Web of Science on 15 March, 2023 using the followed strategy: TS=(critical care OR critically ill OR intensive care OR ICU OR high dependency) AND TS=(deep learning OR convolutional neural network). It should be noted that TS=Topic. The inclusion criteria were as follows: (a) literature published between 2012 and 2022; (b) articles as the type of literature; and (c) literature published in English. Duplicate publications were excluded. A manual check of the included literature was independently performed by two authors (clinicians).

Full records and cited references of the obtained publications were downloaded in BibTex or txt formats for further analysis. Information on the publications, including title, abstract, key words, country, author, institution, source, count of citations, cited references, and the 2021 IF of the top 10 core journals as well as the H-index of the top 10 most productive authors were recorded. Data extraction were conducted by two independent authors.

Bibliometric analysis was performed by Bibliometrix package in R software (4.2.2), Microsoft Excel 2019, VOSviewer (1.6.18), and CiteSpace (5.8.R3). In the present study, publication trends in the literature were analyzed using Microsoft Excel 2019. A polynomial-fitting curve in Microsoft Excel 2019 was applied to predict the number of future publications. The National collaborative networks, author collaborative networks, institutional collaborative networks, and journal publication trends were constructed using Bibliometrix package. Furthermore, co-occurrence of keywords and co-citation relationship of references were analyzed using VOSviewer. Finally, using CiteSpace to analyze keyword bursts.

A total of 1708 articles on DL in CCM were published in the past 11 years. Overall, there was an overall upward trend in the number of publications (Figure 1A), with 3 in 2012 to 651 in 2022. Notably, research activity peaked in 2017, where 95.67% (1634/1708) articles were published during the past six years.

Figure 1 (A) The number of publications and annual citations over time. (B) Curve fitting of the e total annual growth trend of publications (R2 = 0.9773).

Furthermore, the polynomial-fitting curve suggested that research in this area will continue to grow, with an R2 value of 0.9773 (Figure 1B).

Publications on this topic were contributed by 62 different countries/regions. Table 1 shows the top ten most productive countries/regions. China ranked first with 804 publications, followed by the USA with 420 publications. USA had the highest number of total citations and average citations of all publications. Furthermore, China maintained close ties with USA, Korea, and France, whereas USA had strong cooperative bonds with China, England, and Australia (Figure 2).

Table 1 The Top 10 Publishing Countries/Regions

Figure 2 The international collaboration between countries/regions.

Notes: The different colors of arcs represent different countries/regions, and the larger the arc area, the wider the international cooperation of the country/region. Line thickness between countries/regions reflects the intensity of the closeness.

For the analysis of institutions, 2379 institutions made contributions to this field. The top ten most productive institutions were Chinese Academy of Sciences, Harvard University, University of Chinese, Academy of Sciences, University of California System, Wuhan University, Harvard Medical School, Tsinghua University, Centre National De La Recherche Scientifique CNRS, Shanghai Jiao Tong University, and Harbin Institute of Technology (Table 2). Notably, the top 10 institutions were from the China (n = 6), USA (n = 3), and France (n = 1). Furthermore, University of Chinese Academy of Sciences, Harvard University, and Tsinghua University have more connections to other affiliations (Figure 3).

Table 2 The Top 10 Publishing Affiliations

Figure 3 Collaboration between affiliations.

Notes: Each circle represents an affiliation, and the larger the circle, the wider the cooperative relationship. Affiliations with frequent cooperative relationships are clustered into plates of the same color. Line thickness between affiliations reflects the intensity of the closeness.

Over the last ten years, a total of 6211 authors have made significant contributions to the field. Based on publication counts, Wang Y was the most productive author (n = 37), followed by Liu Y, Li Y, Wang J, Wang L, Zhang J, Zhang Y, Li L, Liu J, and Yang Y (Table 3). Furthermore, Zhang Y with the highest total citations and average citations, whereas Li Y has the highest H-index. Interestingly, the top ten most productive authors are all from China. These findings agree with the total productions for the nations mentioned above, showing that China is leading the way in this area. In addition, Wang Y, Wang X, Wang J and Zhang Y have more connections to other authors (Figure 4).

Table 3 The Top 10 Publishing Authors

Figure 4 Collaboration between authors.

Notes: Each circle represents an author, and the larger the circle, the wider the cooperative relationship. Authors with frequent cooperative relationships are clustered into plates of the same colour. Line thickness between authors reflects the intensity of the closeness.

A total of 260 journals have made contributions to this field. As shown in Table 4, IEEE Access, Scientific Reports, Remote Sensing were the top three. When it came to journal impact, IEEE Transactions on Geoscience and Remote Sensing ranked first, with an IF of 8.125, followed by Journal of Biomedical Informatics (IF = 8.000), and Computers in Biology and Medicine (IF = 6.698). These journals were therefore valuable resources for research in this field. Additionally, over the last five years, the top five most active journals have displayed a sharp rise in the number of annual publications (Figure 5).

Table 4 The Top 10 Most Active Journals

Figure 5 Publications of the top 5 most active journals over time.

A total of 59,659 co-cited references were identified. After setting the minimum number of citations to 30, 52 of them were selected to form the cited reference network, which contained four clusters (Figure 6). Cluster 1 (in red) primarily centered on the model development and validation, cluster 2 (in green) primarily centered on the application of the models, cluster 3 (in blue) and cluster 4 (in yellow) primarily centered on the application of models in the medical field. These results suggest that DL-based model development and applications are the foundation of current research in this area. All literature included in Figure 6 is provided in Supplementary Material 1.

Figure 6 Network visualization map of co-citation references.

Notes: Cluster 1 (in red), cluster 2 (in green), cluster 3 (in blue), cluster 4 (in yellow). The lines between the circles represent the co-citation relationship. The thickness and number of connections between the nodes indicate the strength of links between references.

A total of 4864 keywords were identified. Deep learning, machine learning, and feature extraction were the keywords with the highest frequency (Figure 7A). After setting the minimum number of occurrences to ten, 84 of them were selected to form the keyword network, which contained six clusters (Figure 7A). Cluster 1 (in red) primarily centered on the model development and validation, cluster 2 (in green) and cluster 4 (in yellow) primarily centered on the extraction of clinical characteristics of critically ill patients, cluster 3 (in blue) primarily centered on the prognosis of critically ill patients, cluster 5 (in purple) primarily centered on monitoring changes in the condition of critically ill patients; and cluster 6 (in light blue) primarily centered on big data analysis in CCM. It was worth noting that adaptation models, computed tomography, and electronic medical records were recent emerging hot topics (Figure 7B). These topics offer potential research directions on DL in CCM for the future.

Figure 7 (A) Network map of keywords on DL in CCM. (B) Visualization map of top 15 keywords with the strongest citation bursts.

Notes: Cluster 1 (in red), cluster 2 (in green), cluster 3 (in blue), cluster 4, cluster 5 (in purple), cluster 6 (in light blue). The node size reflects the co-occurrence frequencies and the link indicates the co-occurrence relationship. The thickness of the link is proportional to the number of times two keywords co-occur. The blue bars indicate that the keywords have been published and the red bars indicate citation burstness.

The present study used a bibliometric approach to analyze publications of DL in CCM by exploring the expansion of research interest, publication output, top nations, international cooperation, top institutions, authoritative scholars, preferred journals, keywords, and citation analysis.

Publications on DL in CCM have grown steadily in recent years since the concept of DL was introduced in 2006.19 In 2018, there was a significant increase in interest in DL in CCM, which marked the turning point. Interest in DL in general medicine has been gradually increasing since 2012, but in the field of CCM, it was significantly delayed by six years.20 Indeed, the DL models are not commonly used in CCM daily practice. The reason is that few models have external validation, clinical interpretability and high predictability.21,22 Furthermore, most models are developed in a single institution and are do not perform well when applied to other institutions.23 There are also limited venues to incorporate the models. Ideally, they would be embedded in electronic health record systems, but this is challenging to implement due to the limitations of these systems and the corporate disincentives to do so.24,25 In addition, privacy issues are also one of the challenges faced by the adoption of artificial intelligence in the medical field.26 Based on the above evidence, we therefore believe that the safety and accountability of DL models applied to critically ill patients has not yet fully gained gain enough trust from the people. Since 2016, with the continuous development of DL technologies and the development of DL models that begin to pay attention to multi-center data sources and external validation, the accuracy and clinical adaptability of the models have been strengthened, which may help to establish patient confidence in the DL model.27,28 The emergence of new DL technologies, such as convolutional neural networks, long and short-term memory networks, recurrent neural networks, transformer models, and attention mechanisms, offers previously unheard-of possibilities for disease management, diagnosis, and prediction. In addition, MIMIC and eICU, two large public intensive care databases launched in 2016 and 2018, respectively, became available to researchers.29,30 Especially, the release of the MIMIC III database was a large contributing factor to the development of DL models in CCM. Types of DL models developed based on MIMIC III typically include diagnostic models, disease severity score models, real-time monitoring models, hospital length of stay prediction models, readmission prediction models, survival prediction models, and automated adverse drug reaction reporting models.12,15,3135 Common conditions in the field of CCM that these DL models are applied to manage include sepsis, acute respiratory distress syndrome, acute kidney injury, and cardiovascular disease.3135 The common variables used in these models can be divided into 4 categories: history information, admission information, vital signs, laboratory results, and arterial blood gas analysis.15 However, MIMIC-III is only an extensive single-center database spanning from 2001 to 2012 of electronic medical records of patients admitted to the ICU at Beth Israel Deaconess Medical Center, an academic teaching hospital of Harvard Medical School in Boston, USA.36 Therefore, the establishment of a continuously updated CCM database of multicenter admitted ICU patient data, or even electronic medical records of ICU patients admitted globally, would be more conducive to the promotion of the field of CCM as well as the application of DL models.

It was discovered that high-income nations predominate in DL research on CCM after analyzing the distribution of publications across nations. Notably, the top 30 countries in the world in terms of GDP include these ten most productive countries, suggesting that the number of publications is closely linked to the economic power of each country. This finding is consistent with the bibliometric findings of many other medical disciplines.3739 Furthermore, over 70% of the publications came from the USA and China, indicating that these two nations are the main contributors to the DL in CCM research. Additionally, the highest citation rates are also found in these two nations, though China has a lower average citation rate per article than the USA. Analysis of collaborative networks showed that the USA and China are the countries with the most collaborative network relationships. A stable and adaptable policy is a prerequisite to ensure that international collaborations are successfully achieved. Adequate research funding, a wide range of research collaborators, as well as a significant proportion of visiting scholars all contribute to improved international partnerships. Furthermore, the top 100 universities in the world include seven of the top ten most productive institutions, indicating that the use of DL in CCM has gained the attention of leading universities. Researchers may be encouraged to consider conducting conjoint research or applying for educational programs or visiting scholars with these top institutions in the USA or China.

Notably, current articles on this topic are not in the top-tier clinical journals in this space, such as Critical Care Medicine, JAMA, or the New England Journal. It is likely due to the fact that there is still a gap between the DL models and clinical applicability. It needs to be acknowledged that DL has the advantage of responding to the challenges faced by CCM. DL algorithms and models enable machines to mimic human activities such as vision, hearing, and thinking and can automate and reliably process and analyze health information.7 Therefore, various algorithms and models based on DL have been widely used in the management of common diseases in CCM, including early detection of diseases, severity score estimation, facilitating ICU liberation through early successful extubation, early mobility, and survival prediction.1115 Accumulating evidence suggests that the application of DL technology promotes intelligence in CCM, not only effectively improving the quality of medical care, but also helping to increase the efficiency of clinicians.16 DL implementation will support clinicians in the decision-making processes. Benefits comprehend earlier diagnoses, detection of subclinical deteriorations and generation of new medical knowledge.7 To improve the clinical utility of DL models in CCM, the following challenges need to be addressed. First, DL models that lack external validation automatically move away from clinical applicability. It should be advocated that researchers should develop external validation from a multidimensional perspective to continuously improve the scientific validity of the model and enhance the predictive performance, which in turn will promote the clinical interpretability and applicability of DL models in the CMM field.21 Models developed in a single institution are not always applicable to other institutions. Therefore, constructing models based on multicenter shared data should be advocated to increase the breadth of their applicability.23 If achievable, the creation of a global shared database of electronic medical records in the field of CCM is expected to bring a major breakthrough in this area.36 Furthermore, privacy protection through policies remains the cornerstone of health data, with the addition of special safeguards for personal health data addressed by the new innovative principles of the General Data Protection Regulation.40

Based on the authors keywords in the identified categories, CCM-related DL research mainly focused on the model development and validation, the extraction of clinical characteristics of critically ill patients, the prognosis of critically ill patients, monitoring changes in the condition of critically ill patients, and big data analysis in CCM. Furthermore, the primary disease domains addressed in CCM-related DL research were COVID-19, ARDS, sepsis, cardiac arrest, and acute kidney injury. The common targets of DL algorithms are these common diseases in CCM. It was worth noting that adaptation models, computed tomography, and electronic medical records were recent emerging hot topics. These topics offer potential research directions on DL in CCM for the future. DL has demonstrated potential applications in various areas of CCM. However, the development and implementation of DL in CCM remains challenging. Firstly, the absence of external validation and prospective assessment to confirm the repeatability of DL protocols both limit the utility of DL in clinical practice.41 Secondly, implementing artificial intelligence models in clinical practice may entail high initial costs, which is a significant barrier to implementing artificial intelligence (AI) in low- and middle-income countries.42 Furthermore, the World Health Organization released guidelines on the ethics and management of AI for health in 2021, emphasizing the key role of privacy, transparency, informed consent, and regulation of data protection frameworks.43 Thus, the legal protection of patient privacy also limits the current widespread use of AI in medicine.42

The application of AI is beneficial for the advancement of medicine.44,45 This study provides a systematic review of hot spots and trends in CCM-related DL research, highlights leading countries and institutions, reveals potential partnership networks, and provides insights into the direction of future research. However, limitations should be acknowledged. The COVID-19 pandemic has impacted various industries around the world, including the DL field and the CCM field. From the outset of the COVID-19 pandemic, it was clear that the greatest challenge was the unavailability of fully equipped and staffed ICU beds.46 With the expansion of the Internet, the amount of content on COVID-19 has exploded in the last three years. In addition to fact-based content, a large amount of COVID-19 content is being manipulated, and it leads to people spending more time online and getting more invested in this false content.47 Potentially preventing its spread by using DL to identify uninformative information early has also raised concerns.47 Therefore, the COVID-19 pandemic may introduce a publication bias to the publication trend of CCM-related DL research, which may lead to unstable research hotspots and trends identified in this study. Furthermore, given that our search strategy was constructed based on Topic, the search strategy used in the present study may have resulted in missing some relevant literature.

Hot spots in research on the application of DL in CCM have focused on classifying disease phenotypes, predicting early signs of clinical deterioration, and forecasting disease progression, prognosis, and death. Extensive collaborative research to improve the maturity and robustness of the model remains necessary to make DL-based model applications sufficiently compelling for conventional CCM practice.

This work was supported by Science and Technology Development Fund of Hospital of Chengdu University of Traditional Chinese Medicine (No.21ZL08).

The authors report no conflicts of interest in this work.

1. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. Commun ACM. 2017;60:8490. doi:10.1145/3065386

2. Girshick R, Donahue J, Darrell T, Malik J. Rich feature hierarchies for accurate object detection and semantic segmentation. 27th IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Columbus, OH; 2014.

3. Goecks J, Jalili V, Heiser LM, Gray JW. How machine learning will transform biomedicine. Cell. 2020;181(1):92101. doi:10.1016/j.cell.2020.03.022

4. Shinde P, Shah S. A review of machine learning and deep learning applications; 2019.

5. Irisson JO, Ayata SD, Lindsay DJ, Karp-Boss L, Stemmann L. Machine learning for the study of plankton and marine snow from images. Ann Rev Mar Sci. 2022;14:277301. doi:10.1146/annurev-marine-041921-013023

6. Wang S, Yang DM, Rong R, Zhan X, Xiao G. Pathology image analysis using segmentation deep learning algorithms. Am J Pathol. 2019;189(9):16861698. doi:10.1016/j.ajpath.2019.05.007

7. Egger J, Gsaxner C, Pepe A, et al. Medical deep learning-A systematic meta-review. Comput Methods Programs Biomed. 2022;221:106874. doi:10.1016/j.cmpb.2022.106874

8. Zhang Z, Pan Q, Ge H, Xing L, Hong Y, Chen P. Deep learning-based clustering robustly identified two classes of sepsis with both prognostic and predictive values. EBioMedicine. 2020;62:103081. doi:10.1016/j.ebiom.2020.103081

9. Reamaroon N, Sjoding MW, Gryak J, Athey BD, Najarian K, Derksen H. Automated detection of acute respiratory distress syndrome from chest X-rays using directionality measure and deep learning features. Comput Biol Med. 2021;134:104463. doi:10.1016/j.compbiomed.2021.104463

10. Sharma N, Simmons LH, Jones PS, et al. Motor imagery after subcortical stroke: a functional magnetic resonance imaging study. Stroke. 2009;40(4):13151324. doi:10.1161/STROKEAHA.108.525766

11. Lauritsen SM, Kalr ME, Kongsgaard EL, et al. Early detection of sepsis utilizing deep learning on electronic health record event sequences. Artif Intell Med. 2020;104:101820. doi:10.1016/j.artmed.2020.101820

12. Aurolu T, Oul H. A deep learning approach for sepsis monitoring via severity score estimation. Comput Methods Programs Biomed. 2021;198:105816. doi:10.1016/j.cmpb.2020.105816

13. Jia Y, Kaul C, Lawton T, Murray-Smith R, Habli I. Prediction of weaning from mechanical ventilation using convolutional neural networks. Artif Intell Med. 2021;117:102087. doi:10.1016/j.artmed.2021.102087

14. Yeung S, Rinaldo F, Jopling J, et al. A computer vision system for deep learning-based detection of patient mobilization activities in the ICU. NPJ Digit Med. 2019;2:11. doi:10.1038/s41746-019-0087-z

15. Tang H, Jin Z, Deng J, et al. Development and validation of a deep learning model to predict the survival of patients in ICU. J Am Med Inform Assoc. 2022;29(9):15671576. doi:10.1093/jamia/ocac098

16. Datta R, Singh S. Artificial intelligence in critical care: its about time! Med J Armed Forces India. 2021;77(3):266275. doi:10.1016/j.mjafi.2020.10.005

17. Dan F, Kudu E. The evolution of cardiopulmonary resuscitation: global productivity and publication trends. Am J Emerg Med. 2022;54:151164. doi:10.1016/j.ajem.2022.01.071

18. Niu B, Hong S, Yuan J, Peng S, Wang Z, Zhang X. Global trends in sediment-related research in earth science during 19922011: a bibliometric analysis. Scientometrics. 2013;98(1):511529. doi:10.1007/s11192-013-1065-x

19. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521:436444. doi:10.1038/nature14539

20. Guo Y, Hao Z, Zhao S, Gong J, Yang F. Artificial intelligence in health care: bibliometric analysis. J Med Internet Res. 2020;22(7):e18228. doi:10.2196/18228

21. Ozrazgat-Baslanti T, Loftus TJ, Ren Y, Ruppert MM, Bihorac A. Advances in artificial intelligence and deep learning systems in ICU-related acute kidney injury. Curr Opin Crit Care. 2021;27(6):560572. doi:10.1097/MCC.0000000000000887

22. Ching T, Himmelstein DS, Beaulieu-Jones BK, et al. Opportunities and obstacles for deep learning in biology and medicine. J R Soc Interface. 2018;15(141):20170387.

23. Miotto R, Wang F, Wang S, Jiang X, Dudley JT. Deep learning for healthcare: review, opportunities and challenges. Brief Bioinform. 2018;19(6):12361246. doi:10.1093/bib/bbx044

24. Wang L, Tong L, Davis D, Arnold T, Esposito T. The application of unsupervised deep learning in predictive models using electronic health records. BMC Med Res Methodol. 2020;20(1):37. doi:10.1186/s12874-020-00923-1

25. Golas SB, Shibahara T, Agboola S, et al. A machine learning model to predict the risk of 30-day readmissions in patients with heart failure: a retrospective analysis of electronic medical records data. BMC Med Inform Decis Mak. 2018;18(1):44. doi:10.1186/s12911-018-0620-z

26. Harvey HB, Gowda V. Regulatory issues and challenges to artificial intelligence adoption. Radiol Clin North Am. 2021;59(6):10751083. doi:10.1016/j.rcl.2021.07.007

27. Zhong L, Dong D, Fang X, et al. A deep learning-based radiomic nomogram for prognosis and treatment decision in advanced nasopharyngeal carcinoma: a multicentre study. EBioMedicine. 2021;70:103522. doi:10.1016/j.ebiom.2021.103522

28. Hiremath A, Shiradkar R, Fu P, et al. An integrated nomogram combining deep learning, Prostate Imaging-Reporting and Data System (PI-RADS) scoring, and clinical variables for identification of clinically significant prostate cancer on biparametric MRI: a retrospective multicentre study. Lancet Digit Health. 2021;3(7):e445e454. doi:10.1016/S2589-7500(21)00082-0

29. Pollard TJ, Johnson AE, Raffa JD, Celi LA, Mark RG, Badawi O. The eICU collaborative research database, a freely available multi-center database for critical care research. Sci Data. 2018;5:180178. doi:10.1038/sdata.2018.178

30. Johnson AE, Pollard TJ, Shen L, et al. MIMIC-III, a freely accessible critical care database. Sci Data. 2016;3:160035. doi:10.1038/sdata.2016.35

31. Guo F, Zhu X, Wu Z, Zhu L, Wu J, Zhang F. Clinical applications of machine learning in the survival prediction and classification of sepsis: coagulation and heparin usage matter. J Transl Med. 2022;20(1):265. doi:10.1186/s12967-022-03469-6

32. Alfieri F, Ancona A, Tripepi G, et al. External validation of a deep-learning model to predict severe acute kidney injury based on urine output changes in critically ill patients. J Nephrol. 2022;35(8):20472056. doi:10.1007/s40620-022-01335-8

33. Wu J, Lin Y, Li P, Hu Y, Zhang L, Kong G. Predicting Prolonged Length of ICU stay through machine learning. Diagnostics. 2021;11(12):2242. doi:10.3390/diagnostics11122242

34. Pishgar M, Theis J, Del Rios M, Ardati A, Anahideh H, Darabi H. Prediction of unplanned 30-day readmission for ICU patients with heart failure. BMC Med Inform Decis Mak. 2022;22(1):117. doi:10.1186/s12911-022-01857-y

35. McMaster C, Chan J, Liew DFL, et al. Developing a deep learning natural language processing algorithm for automated reporting of adverse drug reactions. J Biomed Inform. 2023;137:104265. doi:10.1016/j.jbi.2022.104265

36. Rsli E, Bozkurt S, Hernandez-Boussard T. Peeking into a black box, the fairness and generalizability of a MIMIC-III benchmarking model. Sci Data. 2022;9(1):24. doi:10.1038/s41597-021-01110-7

37. Qiang W, Xiao C, Li Z, et al. Impactful publications of critical care medicine research in China: a bibliometric analysis. Front Med. 2022;9:974025. doi:10.3389/fmed.2022.974025

38. Liu YX, Zhu C, Wu ZX, Lu LJ, Yu YT. A bibliometric analysis of the application of artificial intelligence to advance individualized diagnosis and treatment of critical illness. Ann Transl Med. 2022;10(16):854. doi:10.21037/atm-22-913

39. Cui X, Chang Y, Yang C, Cong Z, Wang B, Leng Y. Development and trends in artificial intelligence in critical care medicine: a bibliometric analysis of related research over the period of 20102021. J Pers Med. 2022;13(1):50. doi:10.3390/jpm13010050

40. Castelluccia Claude, le Metayer Daniel, European Parliament. European Parliamentary Research Service. Scientific foresight unit. Understanding algorithmic decision-making: opportunities and challenges; 2019. Available from: https://www.europarl.europa.eu/RegData/etudes/STUD/2019/624261/EPRS_STU. Accessed January 27, 2023.

41. Yoon JH, Pinsky MR, Clermont G. Artificial intelligence in critical care medicine. Crit Care. 2022;26(1):75. doi:10.1186/s13054-022-03915-3

42. Caruso PF, Greco M, Ebm C, Angelotti G, Cecconi M. Implementing artificial intelligence: assessing the cost and benefits of algorithmic decision-making in critical care. Crit Care Clin. 2023;6072(1):1253.

43. Ethics and governance of artificial intelligence for health ethics and governance of artificial intelligence for health 2; 2021. Available from: http://apps.who.int/bookorders. Accessed January 27, 2023.

44. Wu Q, Liu S, Zhang R, et al. ACU&MOX-DATA: a platform for fusion analysis and visual display acupuncture multi-omics heterogeneous data. Acupunct Herbal Med. 2023;3(1):5962. doi:10.1097/HM9.0000000000000051

45. Jiang C, Qu H. In-line spectroscopy combined with multivariate analysis methods for endpoint determination in column chromatographic adsorption processes for herbal medicine. Acupunct Herbal Med. 2022;2(4):253260. doi:10.1097/HM9.0000000000000035

46. Arabi YM, Myatra SN, Lobo SM. Surging ICU during COVID-19 pandemic: an overview. Curr Opin Crit Care. 2022;28(6):638644. doi:10.1097/MCC.0000000000001001

47. Machov K, Mach M, Porezan M. Deep learning in the detection of disinformation about COVID-19 in online space. Sensors. 2022;22(23):9319. doi:10.3390/s22239319

Read the rest here:
Research hotspots of deep learning in critical care medicine | JMDH - Dove Medical Press

Protect AI raises $35M to expand its AI and machine learning security platform – VentureBeat

Head over to our on-demand library to view sessions from VB Transform 2023. Register Here

Protect AI, an AI and machine learning (ML) security company, announced it has successfully raised $35 million in a series A funding round. Evolution Equity Partners led the round and saw participation fromSalesforce Venturesand existing investors Acrew Capital, boldstart ventures, Knollwood Capital and Pelion Ventures.

Founded by Ian Swanson, who previously led Amazon Web Services worldwide AI and ML business, the company aims to strengthen ML systems and AI applications against security vulnerabilities, data breaches and emerging threats.

The AI/ML security challenge has become increasingly complex for companies striving to maintain comprehensive inventories of assets and elements in their ML systems. The rapid growth of supply chain assets, such as foundational models and external third-party training datasets, amplifies this difficulty.

These security challenges expose organizations to risks around regulatory compliance, PII leakages, data manipulation and model poisoning.

VB Transform 2023 On-Demand

Did you miss a session from VB Transform 2023? Register to access the on-demand library for all of our featured sessions.

To address these concerns, Protect AI has developed a security platform, AI Radar, that provides AI developers, ML engineers and AppSec professionals real-time visibility, detection and management capabilities for their ML environments.

Machine learning models and AI applications are typically built using an assortment of open-source libraries, foundational models and third-party datasets. AI Radar creates an immutable record to track all these components used in an ML model or AI application in the form of a machine learning bill of materials (MLBOM), Ian Swanson, CEO and cofounder of Protect AI, told VentureBeat. It then implements continuous security checks that can find and remediate vulnerabilities.

>>Dont miss our special issue: The Future of the data center: Handling greater and greater demands.<<

Having secured total funding of $48.5 million to date, the company intends to use the newly acquired funds to scale sales and marketing efforts, enhance go-to-market activities, invest in research and development and strengthen customer success initiatives.

As part of the funding deal, Richard Seewald, founder and managing partner at Evolution Equity Partners, will join the Protect AI board of directors.

The company claims that traditional security tools lack the necessary visibility to monitor dynamic ML systems and data workflows, leaving organizations ill-equipped to detect threats and vulnerabilities in the ML supply chain.

To mitigate this concern, AI Radar incorporates continuously integrated security checks to safeguard ML environments against active data leakages, model vulnerabilities and other AI security risks.

The platform uses integrated model scanning tools for LLMs and other ML inference workloads to detect security policy violations, model vulnerabilities and malicious code injection attacks. Additionally, AI Radar can integrate with third-party AppSec and CI/CD orchestration tools and model robustness frameworks.

The company stated that the platforms visualization layer provides real-time insights into an ML systems attack surface. It also automatically generates and updates a secure, dynamic MLBOM that tracks all components and dependencies within the ML system.

Protect AI emphasizes that this approach guarantees comprehensive visibility and auditability in the AI/ML supply chain. The system maintains immutable time-stamped records, capturing any policy violations and changes made.

AI Radar employs a code-first approach, allowing customers to enable their ML pipeline and CI/CD system to collect metadata during every pipeline execution. As a result, it creates an MLBOM containing comprehensive details about the data, model artifacts and code utilized in ML models and AI applications, explained Protect AIs Swanson. Each time the pipeline runs, a version of the MLBOM is captured, enabling real-time querying and implementation of policies to assess vulnerabilities, PII leakages, model poisoning, infrastructure risks and regulatory compliance.

Regarding the platforms MLBOM compared to a traditional software bill of materials (SBOM), Swanson highlighted that while an SBOM constitutes a complete inventory of a codebase, an MLBOM encompasses a comprehensive inventory of data, model artifacts and code.

The components of an MLBOM can include the data that was used in training, testing and validating an ML model, how the model was tuned, the features in the model, model package formatting, OSS supply chain artifacts and much more, explained Swanson. Unlike SBOM, our platform provides a list of all components and dependencies in an ML system so that users have full provenance of their AI/ML models.

Swanson pointed out that numerous large enterprises use multiple ML software vendors such as Amazon Sagemaker, Azure Machine Learning and Dataiku resulting in various configurations of their ML pipelines.

In contrast, he highlighted that AI Radar remains vendor-agnostic and seamlessly integrates all these diverse ML systems, creating a unified abstraction or single pane of glass. Through this, customers can readily access crucial information about any ML models location and origin and the data and components employed in its creation.

Swanson said that the platform also aggregates metadata on users machine learning usage and workloads across all organizational environments.

The metadata collected can be used to create policies, deliver model BoMs (bills of materials) to stakeholders, and to identify the impact and remediate risk of any component in your ML ecosystem over every platform in use, he told VentureBeat. The solution dashboards user roles/permissions that bridge the gap between ML builder teams and app security professionals.

Swanson told VentureBeat that the company plans to maintain R&D investment in three crucial areas: enhancing AI Radars capabilities, expanding research to identify and report additional critical vulnerabilities in the ML supply chain of both open-source and vendor offerings, and furthering investments in the companys open-source projectsNB DefenseandRebuff AI.

A successful AI deployment, he pointe dout, can swiftly enhance company value through innovation, improved customer experience and increased efficiency.Hence, safeguarding AI in proportion to the value it generates becomes paramount.

We aim to educate the industry about the distinctions between typical application security and security of ML systems and AI applications. Simultaneously, we deliver easy-to-deploy solutions that ensure the security of the entire ML development lifecycle, said Swanson. Our focus lies in providing practical threat solutions, and we have introduced the industrys first ML bill of materials (MLBOM) to identify and address risks in the ML supply chain.

VentureBeat's mission is to be a digital town square for technical decision-makers to gain knowledge about transformative enterprise technology and transact. Discover our Briefings.

Read the original:
Protect AI raises $35M to expand its AI and machine learning security platform - VentureBeat

Machine learning helps this company deliver a better online shopping experience – ZDNet

Westend61/Getty Images

When you're buying a bulky item like a sofa on the internet, the last thing you need is for the product to arrive and for it to look nothing like you expected -- that kind of error is costly and frustrating not only for the person buying the sofa, but also for the company selling the item.

So, what if there was a way to use technology to make the buying experience more enjoyable and to reduce the risk of an unpleasant surprise on delivery day?

Also: This retailer is using RFID tags to make in-person clothes shopping less frustrating

That's exactly what Wayfair is doing with the help of machine learning. The e-commerce company, which sells furniture and home goods online, is using a specially designed platform that's been built alongside technology specialist Snorkel AI.

Tulia Plumettaz, director of machine learning at Wayfair, says the platform is helping her company boost the quality of the online search experience it provides to consumers and to help ensure the sofa you receive looks like the sofa you ordered.

"We have these bulky items that are hard to transport," she says. "We want you to get inspired and feel confident that what you're going to get is what you're buying. And we want that to happen without you even touching the product."

Also: Is Temu legit? What to know about this shopping app before placing your first order

Delivering this kind of high-quality online experience is far from straightforward. Wayfair's site includes thousands of products with a huge number of potential variables, including size, color, and texture.

An added complication comes from the fact that the e-commerce company provides a platform for its suppliers to sell goods to customers. Plumettaz says Wayfair sometimes receives a limited amount of information on products from its suppliers, so providing detailed descriptions to customers can be tough.

It's at that point that Snorkel's platform plays a key role in providing enriched product information.

"We want suppliers to find that it's easy to work with us using our advanced technology. We want them to say, 'I gave Wayfair a picture, some information, and -- with not a lot of effort -- my item just started selling,'" Plumettaz says.

Plumettaz also says machine learning supports "fast-labeling operations" through a bespoke solution that's been developed through a design partnership.

Also: The best online thrifting apps

Snorkel already had its key product called Snorkel Flow, which is a data-centric AI platform for automated data labeling, integrated model training, and analysis.

But while Snorkel Flow is focused on text, Wayfair needed a solution that would support the programmatic labeling of images.

Plumettaz says the solution, which was developed over a twelve-month period by the two companies in combination, provides benefits for both companies: Wayfair gets to shape the technology it's using, and Snorkel gets a route into a new and fast-emerging market.

"We engaged together, and the result is a novel development that brings programmatic labeling into computer vision," says Plumettaz.

Also: Here's why everything on Temu is so cheap

Now, with the bespoke technology in place, Wayfair's team can label and re-label products quickly and effectively.

Rather than having to rely on humans to label up to 40 million products manually, automation deals with a lot of the heavy lifting before specialists within the business -- such as category managers -- ensure the right images are served to online shoppers, says Plumettaz. "With programmatic image labeling, we can match products in the catalogue to the items that customers are looking for as new trends emerge."

Machine learning is also a productivity enhancement -- with less time being spent on labeling images, employees can now focus on higher-value activities. "It's making what we do a lot more interesting," she says. "At Wayfair, our employees don't lack activities to do -- think about maintaining such a rich catalogue. So, now we can be more productive. It's helped make our lives easier and our work a lot more cost-effective."

While Wayfair has chosen to work with Snorkel, Plumettaz recognizes there are other technology players who continue to develop their own machine-learning solutions.

Also: I bought four brand-name tech gadgets on Temu for work. Here's how it went

She says each company has its own stack and, in such a fast-developing market, it's tough to know where machine learning goes next. Plumettaz advises other professionals who are looking at emerging technology to make early inroads and build strong partnerships.

"The field is moving so fast," she says. "Five years ago, it was a lot harder to integrate with a vendor in machine learning. Now, the hurdles to get a vendor approved are disappearing fast."

While machine learning can provide a big boon to customer and employee experiences, Plumettaz says professionals shouldn't let emerging technology work in isolation.

Left to its own devices, an automated system might start labeling products wrongly, leading to unhappy customers and what she refers to as "tremendous consequences".

"You can have an amazing model, but the noise that can come your way through a 1% error rate -- such as when a bulky item gets delivered to your home and it's wrong -- is huge."

The lesson for all business leaders is to ensure the human stays in the loop in what remains a nascent area of development.

Also: I bought some off-brand geeky stuff from Temu (and wasn't mad about it)

"It's a journey with a lot of these applications," she says. "Let's automate, but let's still have a layer that is checking that the automation is working."

Plumettaz provides more details about how that process works at Wayfair. "When we're not confident, we put the outputs in front of a human and get some feedback," she says. "I call it the path to automation. It's like a toddler; it's not yet an adult who can run. And that's the framework that we've been using for those kinds of applications."

Another lesson for professionals who are thinking about dabbling in machine learning is to focus on cross-organization integration and processes, especially in terms of how the technology is implemented, used, and exploited.

Plumettaz says the takeaway will be a familiar one for professionals who introduce new systems or systems: Don't just implement technology for the sake of it. "Partnering really closely with business owners and product owners is key," she says. "I think the blocker is less around the technology and more around thinking about machine learning as a business-value driver from the get-go."

See the article here:
Machine learning helps this company deliver a better online shopping experience - ZDNet

Meet FathomNet: An Open-Source Image Database That Uses Artificial Intelligence and Machine Learning Algorithms … – MarkTechPost

The ocean is changing at an unprecedented rate, making it challenging to maintain responsible stewardship while visually monitoring vast amounts of marine data. The amount and rate of the necessary data gathering are outpacing our capacity to process and analyze them quickly as the research community seeks baselines. The lack of data consistency, inadequate formatting, and the desire for significant, labeled datasets have all contributed to the limited success of recent advancements in machine learning, which have enabled quick and more complex visual data analysis.

In order to meet this requirement, several research institutions worked with MBARI to speed up ocean research by utilizing the capabilities of artificial intelligence and machine learning. One such outcome of this partnership is FathomNet, an open-source image database that employs cutting-edge data processing algorithms to standardize and aggregate carefully curated labeled data. The team believes that using artificial intelligence and machine learning will be the only way to speed up critical studies on ocean health and remove the bottleneck for processing underwater imagery. Details regarding the development process behind this new image database can be found in a recent research publication in Scientific Reports journal.

Machine learning has historically transformed the field of automated visual analysis, partly thanks to vast volumes of annotated data. When it comes to terrestrial applications, the benchmark datasets that machine learning and computer vision researchers swarm to are ImageNet and Microsoft COCO. To give researchers a rich, engaging standard for underwater visual analysis, the team created FathomNet. In order to establish a freely accessible, highly maintained underwater image training resource, FathomNet combines images and recordings from many different sources.

Research workers from MBARIs Video Lab carefully annotated data representing nearly 28,000 hours of deep-sea video and more than 1 million deep-sea photos that MBARI gathered during 35 years. About 8.2 million annotations documenting observations of animals, ecosystems, and objects are present in the video library of MBARI. This comprehensive dataset serves as a priceless tool for the institutes researchers and their international collaborations. Over 1,000 hours of video data were gathered by the Exploration Technology Lab of the National Geographic Society from various marine habitats and places across all ocean basins. These recordings have also been used in the cloud-based collaborative analysis platform developed by CVision AI and annotated by experts from the University of Hawaii and OceansTurn.

Additionally, in 2010, the National Oceanic and Atmospheric Administration (NOAA) Ocean Exploration team aboard the NOAA Ship Okeanos Explorer gathered video data using a dual remotely operated vehicle system. In order to annotate gathered videos more extensively, they started funding professional taxonomists in 2015. Initially, they crowdsourced annotations through volunteer participating scientists. A portion of MBARIs dataset, as well as materials from National Geographic and NOAA, are all included in FathomNet.

Since FathomNet is open source, other institutions can readily contribute to it and utilize it in place of more time- and resource-consuming, conventional methods for processing and analyzing visual data. Additionally, MBARI started a pilot initiative to use machine learning models trained on data from FathomNet to analyze video taken by remotely controlled underwater vehicles (ROVs). Using AI algorithms raised the labeling rate tenfold while reducing human effort by 81 percent. Machine-learning algorithms based on FathomNet data may revolutionize ocean exploration and monitoring. One such example includes using robotic vehicles equipped with cameras and enhanced machine learning algorithms for automatic search and monitoring of marine life and other underwater things.

With ongoing contributions, FathomNet currently has 84,454 images that reflect 175,875 localizations from 81 different collections for 2,243 concepts. The dataset will soon have more than 200 million observations after obtaining 1,000 independent observations for more than 200,000 animal species in various positions and imaging settings. Four years ago, the lack of annotated photos prevented machine learning from examining thousands of hours of ocean film. By unlocking discoveries and enabling tools that explorers, scientists, and the general public may utilize to quicken the pace of ocean research, FathomNet, however, turns this vision into a reality.

FathomNet is a fantastic illustration of how collaboration and community science may promote innovations in our understanding of the ocean. The team believes the dataset can aid in accelerating ocean research when understanding the ocean is more crucial than ever, using data from MBARI and the other collaborators as the foundation. The researchers also emphasize their desire for FathomNet to function as a community where ocean aficionados and explorers from all walks of life may share their knowledge and skills. This will act as a springboard to address problems with ocean visual data that otherwise would not have been achievable without widespread participation. In order to speed up the processing of visual data and create a sustainable and healthy ocean, FathomNet is constantly being improved to include more labeled data from the community.

This Article is written as a research summary article by Marktechpost Staff based on the research paper FathomNet: A global imagedatabase for enabling artifcial intelligence in the ocean. All Credit For This Research Goes To Researchers on This Project. Check out the paper, tool and reference article. Also,dont forget to joinour 26k+ ML SubReddit,Discord Channel,andEmail Newsletter, where we share the latest AI research news, cool AI projects, and more.

Khushboo Gupta is a consulting intern at MarktechPost. She is currently pursuing her B.Tech from the Indian Institute of Technology(IIT), Goa. She is passionate about the fields of Machine Learning, Natural Language Processing and Web Development. She enjoys learning more about the technical field by participating in several challenges.

Original post:
Meet FathomNet: An Open-Source Image Database That Uses Artificial Intelligence and Machine Learning Algorithms ... - MarkTechPost

AI and Machine Learning Hold Potential in Fighting Infectious Disease – HealthITAnalytics.com

July 26, 2023 -A new study described that despite the continued threat of infectious diseases on public health, the capabilities of artificial intelligence (AI) and machine learning (ML) can help handle this issue and provide a framework for future pandemics.

Regardless of research and biological advancements, infectious diseases remain an issue. To keep up with the conflict, common methods that are applied include therapies and diagnostics. Often, synthetic biology approaches provide a platform for innovation. Research indicated that synthetic biology is often divided into two development categories: quantitative biological hypotheses and data from experimentation, and the comprehension of the factors such as nucleic acids and peptides, which allow for the control of biology.

According to research, advancements in AI have considered these factors. Given the complexities of biology and infectious disease, there is a high level of potential. Thus, researchers reviewed how the relationship between AI and synthetic biology can battle infectious diseases.

The review described three uses of AI in infectious diseases: anti-infective drug discovery, infection biology, and diagnostics.

Despite the pre-existence of various anti-infective drugs, drug resistance often outmatches their effectiveness. AI and ML can play a large role in developing new drugs by searching small-molecule databases while using training models to define new drugs or apply existing drugs.

The complications of infection biology are extensive, largely due to the activity of bacterial, eukaryotic, and viral pathogens. These factors can affect host responses, and, therefore, the course of infection.

ML models, however, can analyze nucleic acid, protein, and other variables to determine the aspects of hostpathogen interactions and immune responses. Research also indicates they can define genes and interactions between proteins that link to host cell changes, immunogenicity prediction, and other activities.

Also, gene expression optimization and antigen prediction has assisted the development of vaccines and drugs through supervised models.

AI and ML have applications in diagnostics. As prior instances have shown, the speed of infectious disease detection plays a large role in how spreading takes place. However, through AI and ML, researchers can identify infections and foresee drug resistance. This is primarily because of its ability to program elements well and highlight essential information from biomolecular networks.

Regardless of the opportunities and challenges that these methods may pose, they are essential to the future of infectious disease treatment. As the development of AI continues, it is critical to consider a wide range of datasets to avoid bias.

Various research efforts have also showcased the capabilities of AI and how it may advance healthcare.

Research from April 2022, for example, involved the creation of an AI model that uses non-contrast abdominal CT images to analyze factors related to pancreatic health, determining type 2 diabetes risk.

Using hundreds of images and various measurements, researchers defined the factors that correlated with diabetes. Consistent and accurate results allowed researchers to determine this analysis was an effective approach to detecting diabetes.

This study is a step towards the wider use of automated methods to address clinical challenges, said study authors Ronald M. Summers, MD, PhD, and Hima Tallam, an MD and PhD student, in apress release.It may also inform future work investigating the reason for pancreatic changes that occur in patients with diabetes.

Research efforts such as these are integral examples of how AI continues to play a role in healthcare.

See the rest here:
AI and Machine Learning Hold Potential in Fighting Infectious Disease - HealthITAnalytics.com

Application of machine learning techniques to the modeling of … – Nature.com

Nunes, L. J. R., Causer, T. P. & Ciolkosz, D. Biomass for energy: A review on supply chain management models. Renew. Sustain. Energy Rev. 120, 109658 (2020).

Article Google Scholar

Wang, G. et al. A review of recent advances in biomass pyrolysis. Energy Fuels 34, 1555715578 (2020).

Article CAS Google Scholar

Osman, A. I. et al. Conversion of biomass to biofuels and life cycle assessment: A review. Environ. Chem. Lett. 19, 40754118 (2021).

Article CAS Google Scholar

Bakhtyari, A., Makarem, M. A. & Rahimpour, M. R. Bioenergy Systems for the Future 87148 (Woodhead Publishing, 2017).

Book Google Scholar

Testa, M. L. & Tummino, M. L. Lignocellulose biomass as a multifunctional tool for sustainable catalysis and chemicals: An overview. Catalysts 11, 125 (2021).

Article CAS Google Scholar

Lin, C.-Y. & Lu, C. Development perspectives of promising lignocellulose feedstocks for production of advanced generation biofuels: A review. Renew. Sustain. Energy Rev. 136, 110445 (2021).

Article CAS Google Scholar

Wang, C. et al. A review of conversion of lignocellulose biomass to liquid transport fuels by integrated refining strategies. Fuel Process. Technol. 208, 106485 (2020).

Article CAS Google Scholar

Yamaguchi, A., Sato, O., Mimura, N. & Shirai, M. Catalytic production of sugar alcohols from lignocellulosic biomass. Catal. Today 265, 199202. https://doi.org/10.1016/j.cattod.2015.08.026 (2016).

Article CAS Google Scholar

Erian, A. M. & Sauer, M. Utilizing yeasts for the conversion of renewable feedstocks to sugar alcohols: A review. Bioresour. Technol. 346, 126296. https://doi.org/10.1016/j.biortech.2021.126296 (2022).

Article CAS PubMed Google Scholar

da Costa Lopes, A. M., Joo, K. G., Morais, A. R. C., Bogel-ukasik, E. & Bogel-ukasik, R. Ionic liquids as a tool for lignocellulosic biomass fractionation. Sustain. Chem. Process. 1, 131 (2013).

Article Google Scholar

Abbasi, A. R. et al. Recent advances in producing sugar alcohols and functional sugars by engineering Yarrowia lipolytica. Front. Bioeng. Biotechnol. 9, 648382 (2021).

Article PubMed PubMed Central Google Scholar

Fickers, P., Cheng, H. & SzeKiLin, C. Sugar alcohols and organic acids synthesis in Yarrowia lipolytica: Where are we?. Microorganisms 8, 574 (2020).

Article CAS PubMed PubMed Central Google Scholar

Park, Y.-C., Oh, E. J., Jo, J.-H., Jin, Y.-S. & Seo, J.-H. Recent advances in biological production of sugar alcohols. Curr. Opin. Biotechnol. 37, 105113 (2016).

Article CAS PubMed Google Scholar

Grembecka, M. Sugar alcoholstheir role in the modern world of sweeteners: A review. Eur. Food Res. Technol. 241, 114 (2015).

Article CAS Google Scholar

Amarasekara, A. S. Ionic liquids in biomass processing. Isr. J. Chem. 59, 789802 (2019).

Article CAS Google Scholar

Tan, S. S. Y. & MacFarlane, D. R. Ionic liquids in biomass processing. Ionic Liquids 1, 311339 (2009).

Article Google Scholar

Rajamani, S., Santhosh, R., Raghunath, R. & Jadhav, S. A. Value-added chemicals from sugarcane bagasse using ionic liquids. Chem. Pap. 75, 56055622 (2021).

Article CAS Google Scholar

Parvaneh, K., Rasoolzadeh, A. & Shariati, A. Modeling the phase behavior of refrigerants with ionic liquids using the QC-PC-SAFT equation of state. J. Mol. Liq. 274, 497504. https://doi.org/10.1016/j.molliq.2018.10.116 (2019).

Article CAS Google Scholar

Singh, S. K. & Savoy, A. W. Ionic liquids synthesis and applications: An overview. J. Mol. Liq. 297, 112038 (2020).

Article CAS Google Scholar

Sedghamiz, M. A., Rasoolzadeh, A. & Rahimpour, M. R. The ability of artificial neural network in prediction of the acid gases solubility in different ionic liquids. J. CO2 Util. 9, 3947. https://doi.org/10.1016/j.jcou.2014.12.003 (2015).

Article CAS Google Scholar

Rasoolzadeh, A. et al. A thermodynamic framework for determination of gas hydrate stability conditions and water activity in ionic liquid aqueous solution. J. Mol. Liq. 347, 118358 (2022).

Article CAS Google Scholar

Setiawan, R., Daneshfar, R., Rezvanjou, O., Ashoori, S. & Naseri, M. Surface tension of binary mixtures containing environmentally friendly ionic liquids: Insights from artificial intelligence. Environ. Dev. Sustain. 23, 1760617627 (2021).

Article Google Scholar

Rasoolzadeh, A., Javanmardi, J., Eslamimanesh, A. & Mohammadi, A. H. Experimental study and modeling of methane hydrate formation induction time in the presence of ionic liquids. J. Mol. Liq. 221, 149155. https://doi.org/10.1016/j.molliq.2016.05.016 (2016).

Article CAS Google Scholar

Welton, T. Ionic liquids: A brief history. Biophys. Rev. 10, 691706 (2018).

Article CAS PubMed PubMed Central Google Scholar

Brandt, A., Grsvik, J., Hallett, J. P. & Welton, T. Deconstruction of lignocellulosic biomass with ionic liquids. Green Chem. 15, 550583 (2013).

Article CAS Google Scholar

Reddy, P. A critical review of ionic liquids for the pretreatment of lignocellulosic biomass. S. Afr. J. Sci. 111, 19 (2015).

Article Google Scholar

Tu, W.-C. & Hallett, J. P. Recent advances in the pretreatment of lignocellulosic biomass. Curr. Opin. Green Sustain. Chem. 20, 1117 (2019).

Article Google Scholar

Usmani, Z. et al. Ionic liquid based pretreatment of lignocellulosic biomass for enhanced bioconversion. Biores. Technol. 304, 123003 (2020).

Article CAS Google Scholar

Roy, S. & Chundawat, S. P. S. Ionic liquid-based pretreatment of lignocellulosic biomass for bioconversion: A critical review. BioEnergy Res. 1, 116 (2022).

Google Scholar

Xia, Z. et al. Processing and valorization of cellulose, lignin and lignocellulose using ionic liquids. J. Bioresour. Bioprod. 5, 7995 (2020).

Article CAS Google Scholar

Carneiro, A. P., Rodrguez, O. & Macedo, E. A. Solubility of monosaccharides in ionic liquids: Experimental data and modeling. Fluid Phase Equilib. 314, 2228 (2012).

Article CAS Google Scholar

Carneiro, A. P., Rodrguez, O. & Macedo, E. A. Solubility of xylitol and sorbitol in ionic liquids: Experimental data and modeling. J. Chem. Thermodyn. 55, 184192 (2012).

Article CAS Google Scholar

Carneiro, A. P., Held, C., Rodriguez, O., Sadowski, G. & Macedo, E. A. Solubility of sugars and sugar alcohols in ionic liquids: Measurement and PC-SAFT modeling. J. Phys. Chem. B 117, 99809995 (2013).

Article CAS PubMed Google Scholar

Carneiro, A. P., Rodrguez, O. & Macedo, E. N. A. Fructose and glucose dissolution in ionic liquids: Solubility and thermodynamic modeling. Ind. Eng. Chem. Res. 52, 34243435 (2013).

Article CAS Google Scholar

Mohan, M., Goud, V. V. & Banerjee, T. Solubility of glucose, xylose, fructose and galactose in ionic liquids: Experimental and theoretical studies using a continuum solvation model. Fluid Phase Equilib. 395, 3343 (2015).

Article CAS Google Scholar

Mohan, M., Banerjee, T. & Goud, V. V. Solid liquid equilibrium of cellobiose, sucrose, and maltose monohydrate in ionic liquids: Experimental and quantum chemical insights. J. Chem. Eng. Data 61, 29232932 (2016).

Article CAS Google Scholar

Paduszynski, K., Okuniewski, M. & Domanska, U. Sweet-in-green systems based on sugars and ionic liquids: New solubility data and thermodynamic analysis. Ind. Eng. Chem. Res. 52, 1848218491 (2013).

Article CAS Google Scholar

Paduszyski, K., Okuniewski, M. & Domaska, U. Solidliquid phase equilibria in binary mixtures of functionalized ionic liquids with sugar alcohols: New experimental data and modelling. Fluid Phase Equilib. 403, 167175 (2015).

Article Google Scholar

Paduszyski, K., Okuniewski, M. & Domaska, U. An effect of cation functionalization on thermophysical properties of ionic liquids and solubility of glucose in themmeasurements and PC-SAFT calculations. J. Chem. Thermodyn. 92, 8190 (2016).

Article Google Scholar

Teles, A. R. R. et al. Solubility and solvation of monosaccharides in ionic liquids. Phys. Chem. Chem. Phys. 18, 1972219730 (2016).

Article CAS PubMed PubMed Central Google Scholar

Yang, X., Wang, J. & Fang, Y. Solubility and solution thermodynamics of glucose and fructose in three asymmetrical dicationic ionic liquids from 323.15 K to 353.15 K. J. Chem. Thermodyn. 139, 105879 (2019).

Article CAS Google Scholar

Abbasi, M., Pazuki, G., Raisi, A. & Baghbanbashi, M. Thermophysical and rheological properties of sorbitol+([mmim](MeO)2PO2) ionic liquid solutions: Solubility, density and viscosity. Food Chem. 320, 126566 (2020).

Article CAS PubMed Google Scholar

Zarei, S., Abdolrahimi, S. & Pazuki, G. Thermophysical characterization of sorbitol and 1-ethyl-3-methylimidazolium acetate mixtures. Fluid Phase Equilib. 497, 140150 (2019).

Article Google Scholar

Ruiz-Aceituno, L., Carrero-Carralero, C., Ramos, L. & Sanz, M. L. Selective fractionation of sugar alcohols using ionic liquids. Sep. Purif. Technol. 209, 800805 (2019).

Article CAS Google Scholar

Jeon, P. R. & Lee, C.-H. Artificial neural network modelling for solubility of carbon dioxide in various aqueous solutions from pure water to brine. J. CO2 Util. 47, 101500 (2021).

Article CAS Google Scholar

Amar, M. N. Modeling solubility of sulfur in pure hydrogen sulfide and sour gas mixtures using rigorous machine learning methods. Int. J. Hydrogen Energy 45, 3327433287 (2020).

Article Google Scholar

Hemmati-Sarapardeh, A., Amar, M. N., Soltanian, M. R., Dai, Z. & Zhang, X. Modeling CO2 solubility in water at high pressure and temperature conditions. Energy Fuels 34, 47614776 (2020).

Article CAS Google Scholar

Vanani, M. B., Daneshfar, R. & Khodapanah, E. A novel MLP approach for estimating asphaltene content of crude oil. Pet. Sci. Technol. 37, 22382245 (2019).

Article CAS Google Scholar

Daneshfar, R., Keivanimehr, F., Mohammadi-Khanaposhtani, M. & Baghban, A. A neural computing strategy to estimate dew-point pressure of gas condensate reservoirs. Pet. Sci. Technol. 38, 706712 (2020).

Article CAS Google Scholar

Bakhtyari, A., Mofarahi, M. & Iulianelli, A. Combined mathematical and artificial intelligence modeling of catalytic bio-methanol conversion to dimethyl ether. Energy Convers. Manag. 276, 116562. https://doi.org/10.1016/j.enconman.2022.116562 (2023).

Article CAS Google Scholar

Continued here:
Application of machine learning techniques to the modeling of ... - Nature.com

Machine Learning Engineer Career Path: What You Need to Know – Dice Insights

When it comes to tech skills, machine learning is only getting hotter. Companies of all sizes want tech professionals who can build and manage self-learning A.I. models, and then integrate those models into all kinds of next-generation apps and services.

According tolevels.fyi, which crowdsources compensation information for various tech roles,compensation for those specializing in machine learning and A.I. has increased 2.7 percent over the past six months, from an average of $246,000 per year to $252,535. Those with lots of experience and skill in machine learning can command exponentially higher salaries, of course, especially at big companies known for extremely high pay.

But what does it take to launch yourself onto a machine learning engineer career path, and once youre there, what sort of options are available to you? Lets dive in.

Before embarking on this career path, its important to have a solid foundation in computer science and math, including an understanding of how computers and algorithms work.

Programming is an essential skill, and multiple coding languages may be required, depending on the role and company. Python and JavaScript are often the most popular programming languages that aspiring machine learning engineers focus on first, followed by supporting frameworks like TensorFlow and PyTorch.

Once you've built that foundation and developed your core skill sets, the next step is to start applying what youve learned, explains Neil Ouellette, senior machine learning engineer at Skillsoft. Its important to gain hands-on practice and experience by experimenting with different algorithms and creating small projects on Github.

This is a good way to not only sharpen your skills, but to build a portfolio of work that you can eventually share with prospective employers, he explains.

Do you need formal education to become a machine learning engineer? Thats a great question. Given the demand for ML and AI engineers, many companies are willing to hire tech professionals who dont have a formal two- or four-year degree, provided they can prove during the interview process that they have the skills necessary to succeed in the role. Before you begin applying for jobs, make sure you have a solid grasp on the following, which pop up frequently as requirements for machine learning engineer roles:

In order to carry out these tasks, youll need to have mastered the following:

Mehreen Tahir, software engineer at New Relic, says that entry-level machine learning engineers are often responsible for preprocessing and cleaning data, implementing and testing different machine learning models, and possibly deploying these models.

This involves a lot of data wrangling and debugging, but it's an essential part of the learning process, he says. I always recommend beginners to start working on their own projects or participate in online competitions like those on Kaggle.

These experiences can give you invaluable insights into the practical challenges of machine learning; theyll also help bulk out your resume and application materials when you begin applying for roles in earnest.

An entry-level machine learning engineer (often titled as a junior machine learning engineer or machine learning intern) typically fits into the data science or engineering department of an organization. Some of the typical tasks might include assisting in the development of machine learning models, with lots of collaboration with data analysts and data scientists. As with most tech jobs, a solid grasp of soft skills such as communication and empathy is essential for anyone who wants to make a career out of machine learning.

You might help in building and testing models under the supervision of senior team members, gathering and cleaning data, and learning to interpret and present results, Ouellette says. Many organizations expect their machine learning teams to stay current with the latest techniques and methodologies, and you may be asked to help with this.

There are many pathways for career advancement as a machine learning engineer, whether one is interested in being a manager or individual contributor.

"After gaining some years of experience and expertise, you can advance to a senior role," Ouellette says. "These engineers usually oversee project management, design systems on a larger scale, and may mentor junior engineers."

These potential roles could include a senior machine learning engineer, lead machine learning engineer or team lead, data scientist, AI specialist, machine learning architect, or research scientist. Given the popularity of machine learning, mastering its fundamentals can open an incredible number of career tracks that increasingly rely on the technology.

In the role of team lead, you would oversee and lead a team of machine learning engineers, Ouellette explains. This includes making key decisions on behalf of the team and owning the whole machine learning development process.

In companies that heavily rely on data or A.I., advancing to the executive roles of chief data officer or chief A.I. officer means one is responsible for establishing A.I. and data-related strategies at the highest level.If you find you have a knack for handling clients and translating business problems into data problems, a move into a data science role could be a good fit, Tahir notes.

Data scientists often do a bit of everything, from understanding the business context to data analysis to communicating results in a way that non-technical folks can understand. Soft skills matter more than ever if youre interested in management and want to eventually run your own team.

In this role, you'd be less hands-on with the code and more involved in strategic decisions, team management, and liaising between your team and the rest of the organization, Tahir says. If you're deeply interested in the theoretical side of machine learning and want to push the boundaries of what's possible, you might consider going back to school to get a PhD and become a researcher.

Its important to remember these pathways aren't strictly linear, and the beauty of this field is that there's a lot of flexibility to shape one's own career based on a personal interests and skills. What do you want machine learning to do for you?

To ensure continuous progression in a career as a machine learning engineer, it's crucial to stay updated with the latest advancements in the field, especially as it evolves at a rapid pace. Tools, languages, and frameworks enjoy frequent iterations and updates; if you ignore them for too long, youll fall behind.

Maintaining your baseline knowledge involves taking online courses, attending workshops, webinars, or conferences, and regularly reading relevant research papers. Another key is to constantly work on challenging projects, either at work or in your spare time, that push the boundaries of your current skill set, Tahir says. This hands-on experience is invaluable and can often expose you to new tools and techniques.

Networking is also essential: joining professional groups, online communities, and attending industry events can help you machine learning pros connected, learn from peers, and open new opportunities.

From Tahir's perspective, it's also important to develop soft skills, including communication, teamwork, and problem-solving skills: These are vital, particularly as you move into more senior or managerial roles Demonstrating your ability to effectively communicate complex ideas to non-technical team members or stakeholders can significantly boost your career progression.

Ouellette agrees it's critical to know how to communicate with non-technical audiences. Although machine learning is inherently complex, youll often need to explain how algorithms and statistical models work with stakeholders or clients who may not have a technical background, he says. Strong communication skills are a must.

Go here to read the rest:
Machine Learning Engineer Career Path: What You Need to Know - Dice Insights

Machine learning, incentives and telematics: New tools emerge to … – Utility Dive

The transition to electric vehicles will require significant new amounts of power generation for charging, but utilities say those resources can be developed in time. A more pressing challenge may be managing new charging loads, ensuring millions of vehicles do not put undue stress on the grid.

There will be 30 million to 42 million electric vehicles on U.S. roads in 2030, and they will require about 28 million charging ports,according to the National Renewable Energy Laboratory. Utilities, distributed energy resource aggregators and research institutions are all stepping up to address the issue.

Power generation is only a part of this conversation. Just as important is improving our ability to manage demand in real time, Albert Gore, executive director of the Zero Emission Transportation Association, said Monday in a discussion of how the utility sector must approach EVs.

The industry needs to further its ability to precisely manage demand in real time, including by accurately predicting when and where increases in demand will occur, according to a new ZETA policy brief.

Utilities particularly larger electricity providers in urban areas have been working for years to nudge EV charging to off-peak hours through time-of-use rates or EV-specific rates.

Consolidated Edison, which serves New York City, expects more than a quarter million EVs in its territory by 2025 and has been working since 2017 to encourage grid-beneficial charging through its SmartCharge program, which offers incentives for drivers to avoid charging during peak times.

It's one of, if not the most, successful managed charging programs in the country,Cliff Baratta, Con Edisons electric vehicle strategy and markets section manager, said during ZETAs discussion. At the end of 2022, the utility had 20% of all light-duty EVs registered in its territory enrolled in the program.

In a lot of other places, we see that 5-6% is considered good, Baratta said. We have been able to get really strong engagement with that program, to try and entrench this grid beneficial charging behavior.

Research institutions are working to develop solutions. Argonne National Laboratory and the University of Chicago have partnered on the development of a new algorithm to manage EV charging that utilizes machine learning to efficiently schedule loads.

Distributed energy resource managers are rolling out approaches to managing the anticipated demand..

FlexCharging, which has provided managed charging programs and pilots since 2019, is rolling out a product called EVisionfor smaller utilities that may have fewer resources to devote to demand management initiatives.

Cloud-based software company Virtual Peaker on Tuesday launched a managed charging solution that allows utilities to utilize both vehicle telematics data or internet-connected EV chargers to manage vehicles in charging programs.

The company is focusing on creating a single, scalable solution to increase adoption of distributed energy resources programs and help utilities reach their goals more quickly and efficiently, Virtual Peaker founder and CEO Williams Burke said in a statement.

The companys DER platform is already being used by Efficiency Maine, the states administrator for energy efficiency and demand management programs, to manage battery systems and EV chargers during peak demand periods.

Read the original:
Machine learning, incentives and telematics: New tools emerge to ... - Utility Dive

USC at the International Conference on Machine Learning (ICML … – USC Viterbi School of Engineering

USC researchers will present nine papers at the 40th International Conference on Machine Learning (ICML 2023).

This year, USC researchers will showcase nine papers at the 40th International Conference on Machine Learning (ICML 2023), one of the most prestigious machine learning conferences, taking place July 23 29 in Honolulu, Hawaii. ICML brings together the artificial intelligence (AI) community to share new ideas, tools, and datasets, and make connections to advance the field.

Accepted papers with USC affiliation:

Moccasin: Efficient Tensor Rematerialization for Neural Networks Tu, Jul 25, 17:00Poster Session 2

Burak Bartan,Haoming Li,Harris Teague,Christopher Lott,Bistra Dilkina

Abstract: The deployment and training of neural networks on edge computing devices pose many challenges. The low memory nature of edge devices is often one of the biggest limiting factors encountered in the deployment of large neural network models. Tensor rematerialization or recompute is a way to address high memory requirements for neural network training and inference. In this paper we consider the problem of execution time minimization of compute graphs subject to a memory budget. In particular, we develop a new constraint programming formulation called textsc{Moccasin} with onlyO(n)integer variables, wherenis the number of nodes in the compute graph. This is a significant improvement over the works in the recent literature that propose formulations withO(n2)Boolean variables. We present numerical studies that show that our approach is up to an order of magnitude faster than recent work especially for large-scale graphs.

Refined Regret for Adversarial MDPs with Linear Function Approximation Tu, Jul 25, 17:00 Poster Session 2

Yan Dai,Haipeng Luo,Chen-Yu Wei,Julian Zimmert

Abstract: We consider learning in an adversarial Markov Decision Process (MDP) where the loss functions can change arbitrarily overKepisodes and the state space can be arbitrarily large. We assume that the Q-function of any policy is linear in some known features, that is, a linear function approximation exists. The best existing regret upper bound for this setting (Luo et al., 2021) is of order(K2/3)(omitting all other dependencies), given access to a simulator. This paper provides two algorithms that improve the regret to(K)in the same setting. Our first algorithm makes use of a refined analysis of the Follow-the-Regularized-Leader (FTRL) algorithm with the log-barrier regularizer. This analysis allows the loss estimators to be arbitrarily negative and might be of independent interest. Our second algorithm develops a magnitude-reduced loss estimator, further removing the polynomial dependency on the number of actions in the first algorithm and leading to the optimal regret bound (up to logarithmic terms and dependency on the horizon). Moreover, we also extend the first algorithm to simulator-free linear MDPs, which achieves(K8/9)regret and greatly improves over the best existing bound(K14/15). This algorithm relies on a better alternative to the Matrix Geometric Resampling procedure by Neu & Olkhovskaya (2020), which could again be of independent interest.

Searching Large Neighborhoods for Integer Linear Programs with Contrastive Learning We, Jul 26, 14:00 Poster Session 3

Taoan Huang,Aaron Ferber,Yuandong Tian,Bistra Dilkina,Benoit Steiner

Abstract: Integer Linear Programs (ILPs) are powerful tools for modeling and solving a large number of combinatorial optimization problems. Recently, it has been shown that Large Neighborhood Search (LNS), as a heuristic algorithm, can find high-quality solutions to ILPs faster than Branch and Bound. However, how to find the right heuristics to maximize the performance of LNS remains an open problem. In this paper, we propose a novel approach, CL-LNS, that delivers state-of-the-art anytime performance on several ILP benchmarks measured by metrics including the primal gap, the primal integral, survival rates and the best-performing rate. Specifically, CL-LNS collects positive and negative solution samples from an expert heuristic that is slow to compute and learns a new one with a contrastive loss. We use graph attention networks and a richer set of features to further improve its performance.

Fairness in Matching under UncertaintyWe, Jul 26, 17:00Poster Session 4

Siddartha Devic,David Kempe,Vatsal Sharan,Aleksandra Korolova

Abstract: The prevalence and importance of algorithmic two-sided marketplaces has drawn attention to the issue of fairness in such settings. Algorithmic decisions are used in assigning students to schools, users to advertisers, and applicants to job interviews. These decisions should heed the preferences of individuals, and simultaneously be fair with respect to their merits (synonymous with fit, future performance, or need). Merits conditioned on observable features are always emph{uncertain}, a fact that is exacerbated by the widespread use of machine learning algorithms to infer merit from the observables. As our key contribution, we carefully axiomatize a notion of individual fairness in the two-sided marketplace setting which respects the uncertainty in the merits; indeed, it simultaneously recognizes uncertainty as the primary potential cause of unfairness and an approach to address it. We design a linear programming framework to find fair utility-maximizing distributions over allocations, and we show that the linear program is robust to perturbations in the estimated parameters of the uncertain merit distributions, a key property in combining the approach with machine learning techniques.

On Distribution Dependent Sub-Logarithmic Query Time of Learned Indexing Th, Jul 27, 13:30Poster Session 5

Sepanta Zeighami,Cyrus Shahabi

Abstract: A fundamental problem in data management is to find the elements in an array that match a query. Recently, learned indexes are being extensively used to solve this problem, where they learn a model to predict the location of the items in the array. They are empirically shown to outperform non-learned methods (e.g., B-trees or binary search that answer queries inO(logn)time) by orders of magnitude. However, success of learned indexes has not been theoretically justified. Only existing attempt shows the same query time ofO(logn), but with a constant factor improvement in space complexity over non-learned methods, under some assumptions on data distribution. In this paper, we significantly strengthen this result, showing that under mild assumptions on data distribution, and the same space complexity as non-learned methods, learned indexes can answer queries inO(loglogn)expected query time. We also show that allowing for slightly larger but still near-linear space overhead, a learned index can achieveO(1)expected query time. Our results theoretically prove learned indexes are orders of magnitude faster than non-learned methods, theoretically grounding their empirical success.

SurCo: Learning Linear SURrogates for COmbinatorial Nonlinear Optimization ProblemsTh, Jul 27, 16:30Poster Session 6

Aaron Ferber,Taoan Huang,Daochen Zha,Martin Schubert,Benoit Steiner,Bistra Dilkina,Yuandong Tian

Abstract: Optimization problems with nonlinear cost functions and combinatorial constraints appear in many real-world applications but remain challenging to solve efficiently compared to their linear counterparts. To bridge this gap, we proposeSurCothat learns linearSurrogate costs which can be used in existingCombinatorial solvers to output good solutions to the original nonlinear combinatorial optimization problem. The surrogate costs are learned end-to-end with nonlinear loss by differentiating through the linear surrogate solver, combining the flexibility of gradient-based methods with the structure of linear combinatorial optimization. We propose threevariants:for individual nonlinear problems,for problem distributions, andto combine both distribution and problem-specific information. We give theoretical intuition motivating, and evaluate it empirically. Experiments show thatfinds better solutions faster than state-of-the-art and domain expert approaches in real-world optimization problems such as embedding table sharding, inverse photonic design, and nonlinear route planning.

Emergent Asymmetry of Precision and Recall for Measuring Fidelity and Diversity of Generative Models in High Dimensions Th, Jul 27, 13:30Poster Session 5

Mahyar Khayatkhoei,Wael AbdAlmageed

Abstract: Precision and Recall are two prominent metrics of generative performance, which were proposed to separately measure the fidelity and diversity of generative models. Given their central role in comparing and improving generative models, understanding their limitations are crucially important. To that end, in this work, we identify a critical flaw in the common approximation of these metrics using k-nearest-neighbors, namely, that the very interpretations of fidelity and diversity that are assigned to Precision and Recall can fail in high dimensions, resulting in very misleading conclusions. Specifically, we empirically and theoretically show that as the number of dimensions grows, two model distributions with supports at equal point-wise distance from the support of the real distribution, can have vastly different Precision and Recall regardless of their respective distributions, hence an emergent asymmetry in high dimensions. Based on our theoretical insights, we then provide simple yet effective modifications to these metrics to construct symmetric metrics regardless of the number of dimensions. Finally, we provide experiments on real-world datasets to illustrate that the identified flaw is not merely a pathological case, and that our proposed metrics are effective in alleviating its impact.

Conformal Inference is (almost) Free for Neural Networks Trained with Early Stopping Th, Jul 27, 13:30Poster Session 5

Ziyi Liang,Yanfei Zhou,Matteo Sesia

Abstract: Early stopping based on hold-out data is a popular regularization technique designed to mitigate overfitting and increase the predictive accuracy of neural networks. Models trained with early stopping often provide relatively accurate predictions, but they generally still lack precise statistical guarantees unless they are further calibrated using independent hold-out data. This paper addresses the above limitation with conformalized early stopping: a novel method that combines early stopping with conformal calibration while efficiently recycling the same hold-out data. This leads to models that are both accurate and able to provide exact predictive inferences without multiple data splits nor overly conservative adjustments. Practical implementations are developed for different learning tasks outlier detection, multi-class classification, regression and their competitive performance is demonstrated on real data.

A Critical View of Vision-Based Long-Term Dynamics Prediction Under Environment MisalignmentTu, Jul 25, 17:00Poster Session 2

Hanchen Xie,Jiageng Zhu,Mahyar Khayatkhoei,Jiazhi Li,Mohamed Hussein,Wael AbdAlmageed

Abstract: Dynamics prediction, which is the problem of predicting future states of scene objects based on current and prior states, is drawing increasing attention as an instance of learning physics. To solve this problem, Region Proposal Convolutional Interaction Network (RPCIN), a vision-based model, was proposed and achieved state-of-the-art performance in long-term prediction. RPCIN only takes raw images and simple object descriptions, such as the bounding box and segmentation mask of each object, as input. However, despite its success, the models capability can be compromised under conditions of environment misalignment. In this paper, we investigate two challenging conditions for environment misalignment: Cross-Domain and Cross-Context by proposing four datasets that are designed for these challenges: SimB-Border, SimB-Split, BlenB-Border, and BlenB-Split. The datasets cover two domains and two contexts. Using RPCIN as a probe, experiments conducted on the combinations of the proposed datasets reveal potential weaknesses of the vision-based long-term dynamics prediction model. Furthermore, we propose a promising direction to mitigate the Cross-Domain challenge and provide concrete evidence supporting such a direction, which provides dramatic alleviation of the challenge on the proposed datasets.

Published on July 25th, 2023

Last updated on July 26th, 2023

Go here to see the original:
USC at the International Conference on Machine Learning (ICML ... - USC Viterbi School of Engineering