Page 2,172«..1020..2,1712,1722,1732,174..2,1802,190..»

Dask-ML dask-ml 2021.11.31 documentation

Dask-ML provides scalable machine learning in Python using Dask alongsidepopular machine learning libraries like Scikit-Learn, XGBoost, and others.

People may run into scaling challenges along a couple dimensions, and Dask-MLoffers tools for addressing each.

The first kind of scaling challenge comes when from your models growing solarge or complex that it affects your workflow (shown along the vertical axisabove). Under this scaling challenge tasks like model training, prediction, orevaluation steps will (eventually) complete, they just take too long. Youvebecome compute bound.

To address these challenges youd continue to use the collections you know andlove (like the NumPy ndarray, pandas DataFrame, or XGBoost DMatrix)and use a Dask Cluster to parallelize the workload on many machines. Theparallelization can occur through one of our integrations (like Dasksjoblib backend to parallelize Scikit-Learn directly) or one ofDask-MLs estimators (like our hyper-parameter optimizers).

The second type of scaling challenge people face is when their datasets growlarger than RAM (shown along the horizontal axis above). Under this scalingchallenge, even loading the data into NumPy or pandas becomes impossible.

To address these challenges, youd use Dasks one of Dasks high-levelcollections like(Dask Array, Dask DataFrame or Dask Bag) combined with one of Dask-MLsestimators that are designed to work with Dask collections. For example youmight use Dask Array and one of our preprocessing estimators indask_ml.preprocessing, or one of our ensemble methods indask_ml.ensemble.

Its worth emphasizing that not everyone needs scalable machine learning. Toolslike sampling can be effective. Always plot your learning curve.

In all cases Dask-ML endeavors to provide a single unified interface around thefamiliar NumPy, Pandas, and Scikit-Learn APIs. Users familiar withScikit-Learn should feel at home with Dask-ML.

Other machine learning libraries like XGBoost already havedistributed solutions that work quite well. Dask-ML makes no attempt tore-implement these systems. Instead, Dask-ML makes it easy to use normal Daskworkflows to prepare and set up data, then it deploys XGBoostalongside Dask, and hands the data over.

See Dask-ML + XGBoost for more information.

View original post here:
Dask-ML dask-ml 2021.11.31 documentation

Read More..

How Snapchat Is Using AI And Machine Learning To Thwart Drug Deals – Hot Hardware

Snapchat is taking a proactive approach in fighting drug deals taking place on its social media platform. The company shared an update concerning its most recent efforts to halt the push to sell drugs through connections on its app.Snapchat typically makes the news when the social media platform goes dark, sending users in a frenzy wondering when their beloved app will be back up and running. But the extremely popular app, especially among teenagers and young adults, found itself in a different type of spotlight last October when NBC News did a story about the troubling drug deals presumably taking place on the app.

The report examined the death of teens and young adults who were suspected of buying fentanyl-laced drugs using Snapchat. In the report, it spoke about teens and young adults who had bought what they believed to be a prescription pill, but turned out to be a counterfeit pill containing deadly doses of fentanyl. Since that report, Snapchat has been ramping up its efforts to thwart drug deals on its platform.

Snap stated that it has a zero tolerance for drug dealing on Snapchat. It says it has made significant operational improvements over the past year toward its goal of completely eradicating drug dealers from its platform. It claims to take a holistic approach, which includes "deploying tools that proactively detect drug-related content, working with law enforcement to support their investigations, and provide in-app information and support to Snapchatters who search for drug-related terms through a new educational portal, Heads Up."

The social media company announced that it is adding two new partners to its Heads Up portal in order to provide important in-app resources to its users. Community Anti-Drug Coalitions of America (CADCA), is a nonprofit organization that focuses on creating safe, healthy and drug-free communities. Truth Initiative is the second addition, and is an organization that strives to steer teens and young adults away from smoking, vaping and nicotine in general. Along with these two new organizations being added, Snap will soon be releasing its next episode of it special Good Luck America series which will focus on fentanyl.

Snapchat is also updating its Quick Add suggestion feature in order to reduce interactions between kids and strangers. The company states, "In order to be discoverable in Quick Add by someone else, users under 18 will need to have a certain number of friends in common with that person." In the past, users would be given a list of recommended friends based on mutual connections, regardless if you knew the person in real life or not. Work is also being done on additional parental tools that it will roll out in the coming months.

Another way Snapchat is looking to deter drug dealers from using its platform, is in its cooperation with law enforcement. It has implemented measures using artificial intelligence (AI) and machine learning to identify drug slang and content on the app, and then works with law enforcement to report potential cases and to comply with information requests. Snapchat has increased its law enforcement operations team by 74% since its creation. And remarkably, Snapchat claims that a whopping 88% of drug related content it uncovers is proactively detected by its AI and machine learning algorithms. That's up from 33% since its previous update.

"When we find drug dealing activity, we promptly ban the account, use technology to block the offender from creating new accounts on Snapchat, and in some cases proactively refer the account to law enforcement for investigation," Snapchat says.

Continued here:
How Snapchat Is Using AI And Machine Learning To Thwart Drug Deals - Hot Hardware

Read More..

Application of ensemble machine learning algorithms on lifestyle factors and wearables for cardiovascular risk prediction | Scientific Reports -…

Data source and study population

Data used in this study was drawn from a SingHEART prospective longitudinal cohort study (ClinicalTrials.gov Identifier: NCT02791152). The study is a multi-ethnic population-based study conducted on healthy Asians, aged 2169years old without known diabetes mellitus or prior cardiovascular disease (Ischemic heart disease, stroke, peripheral vascular disease). The study complied with the Declaration of Helsinki and written informed consent were given by participants. The study was approved by the SingHealth Centralized Institutional Review Board.

We included 600 volunteers, aged of 30years with valid calcium score, into the main analysis of this study. Two hundred volunteers under the age of 30years, who did not have a calcium score were excluded, as the calcium score was the main outcome of our analysis.

Subset analysis for activity tracker data was performed on 430 out of the 600 volunteers who had adequate data. Although subjects recruited were issued an activity tracker to be worn over a period of five days with first and last days of the study being partial days, there was inconsistent wearing of the activity. Discounting the partial days, each subject would yield an activity log for three complete tracking days (or equivalent to days with>20 valid hours of steps and sleep data)24,25. For data consistency and quality, subjects with improper activity tracker usage i.e. activity reading log less than five days and/or sleep reading log less than three days were censored.

Coronary artery calcium (CAC) scoring was used as the modelling outcome. The coronary calcium is a specific marker of coronary atherosclerosis, a precursor for coronary artery disease26; it also reflects arterial age under the influence of underlying comorbidities and lifestyle. The CAC score was also regarded as the best marker for risk prediction of cardiovascular events27,28.

This study stratified subjects into two classes of CVD risk. Low risk if their coronary artery calcium score were 0, and high risk if calcium score were 100 and above. Subjects who did not fall into these 2 categories were considered intermediate risk.

The aim of this study is to look at how accurate the machine learning algorithm is in handling different data types, in the task of predicting high risk and low risk patients, based on calcium score.

Table 1 summarizes the data from SingHEART that was used in this study.

Data variables were categorized into four groups; lifestyle survey questionnaires, blood test data, 24-h ambulatory blood pressure, and activity tracking data by commercially available Fitbit Charge HR29.

Data pre-processing, transformation and imputation were performed on the raw data. Variables selected were based on their a priori knowledge from previous publications on cardiovascular risk assessment1,2,3, and physician expert advice. In total, there were 30, 17, 12 and 16 unique variables in the respective groups: survey questionnaire, 24h blood pressure and heart rate monitoring, blood tests and Fitbit data.

The Framingham 10-year risk score was computed using seven traditional risk factors: gender, age, single timepoint systolic blood pressure, Total Cholesterol (TC), High Density Lipoprotein (HDL), smoking status and presence of diabetes. A Framingham risk score of<10% is consider low risk, while20% is considered high risk30.

Figure1 shows the methodological framework of the present study. Exploratory analysis showed that ensemble MLA classifiers were superior at discriminating low risk individuals while ensemble MLA regressors performed better identifying individuals with high CVD risk. To leverage on the merits of both the classifiers and regressors MLA, we used both approaches for our model.

Modelling flow chart using ensemble MLA for cardiovascular risk prediction.

The ensemble classifiers produce a binary prediction outcome; low or non-low risk. The ensemble regressors makes a numerical prediction on the calcium score for individuals classified as non-low risk, and stratify into three bins of low, high, and intermediate risk. The predicted numerical values may range from negative to positive number. Negative predicted values were first converted to zero and subsequently the continuous predictions were converted to discrete bins using unique value percentile discretization ensuring records with the same numerical prediction are assigned the same risk category. Finally, the prediction outcome resides in a decision node build on a rule-based logic. The decision node assigns an outcome of low risk if classifiers predict an individual to be low in CVD risk, high risk if classifier predicts non-low risk and regressor predicts high risk. Patients with incongruent classifiers and regressor outcomes are considered unclassified.

The ensemble models in both classification and regression phase each fit three base learners (naive bayes (NB), random forest (RF) and support vector classifier (SVC) for classification prediction, and generalized linear regression (GLM), support vector regressor (SVR) and stochastic gradient descent (SGD) for regression prediction). These base learners were chosen based on preliminary analysis, where these models showed efficiency in handling missing values and outliers.

The ensemble model then uses majority vote to determine the class label in classification phase. For the regression phase, the ensemble model averages the normalized predictions from the base regressor models to form a numerical outcome.

All models were trained on a stratified five-fold cross-validation. As SingHEART data had an imbalanced CVD risk distribution of risk based on the calcium score (low risk 63.4%, high risk 8.3%, intermediate risk 18.7%) we oversampled the training set for the minority class labels to allow model to better learn features from the under-represented classes31. The data were first partitioned into five mutually exclusive subsets, with each subset sharing the same proportion of class label as original dataset. At each iteration, the MLAs trained on four parts (80%) and validated on the fifth, the holdout set (20%). The process repeats five times, with five different but overlapping training sets. The resulting metrics from each fold were averaged to produce a single estimate.

To simulate access to the different variable groups as per clinical workflow and ease of information availability, we assessed the performance of individual variable group, and in combination as per the following:

Model 1: Survey Questionnaire.

Model 2: 24h ambulatory blood pressure and heart rate.

Model 3: Clinical blood results.

Model 4: Model 1+Model 2.

Model 5: Model 1+Model 3.

Model 6: Model 1 to Model 3.

Model 6*: Model 1 to Model 3 with feature selection.

Model 7: Physical activity and sleep trackers (exploratory subset analysis).

Variables in model 6* were reduced using SVC recursive feature elimination with cross-validation (SVC-RFECV) method to automatically select the best set of predictors that yield the highest area under Receiver Operating Characteristic curves (AUC). Model 16 were trained using 600 subjects.

We also performed exploratory analysis using MLA on the Fitbit Charge HR data (Model 7). Model 7 was trained on a subset of 430 subjects constrained by availability of valid activity tracking data.

Since no single metric can objectively evaluate the cardiovascular risk prediction, we evaluate the performance of our models at CVD risk class level using a panel of metrics; sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), F1-score and Area under Receiver Operating Characteristic curves (AUC). Overall discriminative ability of the model was described by the area under received operating characteristic curve (ROC). All AUC metrics were accompanied by 95% confidence interval (CI) and standard deviation (SD).

To better understand the relative importanceof different risk factors, we conduct a post-hoc approach to rank the variables by their contribution to CVD risk prediction. Feature importance were obtained from the SVC algorithm where the relative importance was determined by the absolute size of the coefficients in relation to others. All statistical analyses were conducted on Python version 3.7 environment and all MLAs and evaluation metrics were constructed using Scikit-learn libraries.

Continue reading here:
Application of ensemble machine learning algorithms on lifestyle factors and wearables for cardiovascular risk prediction | Scientific Reports -...

Read More..

Data Vault Holdings Expands Expertise In Artificial Intelligence, Machine Learning, and Big Data; Appoints Tony Evans of C3 AI To Advisory Board -…

NEW YORK, Jan. 19, 2022 /PRNewswire/ --Data Vault Holdings Inc., the emerging leader in metaverse data visualization, valuation, and monetization announced today the appointment of Tony Evans, General Manager of Financial Services for C3 AI (NYSE: AI), to its advisory board, fortifying Data Vault Holding's expertise in artificial intelligence, machine learning, fintech, e-commerce and security. A preeminent expert in business and sales, Mr. Evans has developed and executed transformative, customer-focused strategies across industries. From artificial intelligence to cybersecurity to e-commerce, he has managed global sales and partnership development, led global banking teams, driven growth, and developed customer big data and innovation strategies. As a member of the advisory board, Mr. Evans will advise Datavaultleadership on the automation and scale of their comprehensive crypto data solution.

"In my role at C3 AI, I witness daily the power data assets and tokenomics can play in the foundation for predictive technology that influences decisions and leads to disruption of incumbent markets. Data has now become both an indicator of business intelligence and a form of capital, and we can use this information to inform business innovation. Datavaultexpertly combines artificial intelligence, machine learning, and crypto-technology to transform data into salable business growth and revenues. I am honored to provide Datavault's leadership with perspective on emerging trends, market impact, and consumer issues in payments, AI, and data," says Tony Evans, General Manager of Financial Services for C3 AI.

As General Manager of Financial Services of leading enterprise AI software provider C3 AI, Mr. Evans directs financial services strategy, global sales, and partnership development. His expertise supports the delivery of the cross-industries enterprise platform C3 AI Suite, which enables businesses to develop, deploy, and operate large-scale AI, predictive analytics, and Internet of Things (IoT) applications. Mr. Evans' diversified background in the financial and technology sectors skillfully positions him to provide counsel to the executive team of Data Vault Holdings, as they develop and launch new products, design new revenue models, and simplify data visualization, valuation, and monetization processes layering effects through automation of their novel crypto-technologies. Additionally, Mr. Evans has served as Leader of Global Banking and Payments and Head of Financial Services (UK) with Amazon Web Services (AWS) (NASDAQ: AMZN); Head of Leonardo and Analytics (UK and Ireland) and SVP & Chief Operating Officer of Financial Services for SAP; and Managing Director (US) of BlackBerry. He has also served in leadership roles with Datawatch Corporation, Oracle, Applied Knowledge LTD, Visusol Consulting, and Smith Industries.

Mr. Evans holds an MBA specializing in business growth, change movement, and change strategy from the University of Brighton.

In coming weeks, Data Vault Holdings plans to announce additional members of its advisory board, with Ed Cushing, Global Account Manager at Amazon Web Services (AWS),recently announced as its inaugural member. New advisory board members will further aid in providing market insights, analytics expertise, and business and data monetization strategies through the use of Datavault's patented, cloud-based SaaS platform.

About Data Vault Holdings Inc.

Data Vault Holdings Inc. is a technology holding company that provides a proprietary, cloud-based platform for the delivery of branded data-backed cryptocurrencies. Data Vault Holdings Inc. provides businesses with the tools to monetize data assets securely over its Information Data Exchange(IDE). The company is in the process of finalizing the consolidation of its affiliates Data Donate Technologies, Inc., ADIO LLC, and Datavault Inc. as wholly owned subsidiaries under one corporate structure. Learn more about Data Vault Holdings Inc. here.

SOURCE Data Vault Holdings Inc.

Originally posted here:
Data Vault Holdings Expands Expertise In Artificial Intelligence, Machine Learning, and Big Data; Appoints Tony Evans of C3 AI To Advisory Board -...

Read More..

The 6 Best Deep Learning Tutorials on YouTube to Watch Right Now – Solutions Review

Learning deep learning can be a complicated process, and its not easy to know where to start. As a result, our editors have compiled this list of the best deep learning tutorials on YouTube to help you learn about the topic and hone your skills before you move on to mastering it. All of the videos here are free to access and feature guidance from some of the top minds and biggest brands in the online learning community. All of the best deep learning tutorials listed tout a minimum of 200,000 views.

Note: Dont forget to subscribe to Solutions Review on YouTube!

Author: Lex Fridman

Description: An introductory lecture for MIT course 6.S094 on the basics of deep learning including a few key ideas, subfields, and the big picture of why neural networks have inspired and energized an entirely new generation of researchers. The most popular result on YouTube touts nearly 1.2 million views.

Author: sentdex

Description: An updated deep learning introduction using Python, TensorFlow, and Keras, this tutorial tours more than a million views and is one of the most popular resources on the web. Students can access the text-tutorial and notes here, TensorFlow Docs here, and Keras docs here. There is also a community Discord server for those interested in the topic.

Author: Simplilearn

Description: This video provides a fun and simple introduction to deep learning concepts. Students learn about where deep learning is implemented and move on to how it is different from machine learning and artificial intelligence. Watchers will also look at what neural networks are and how they are trained to recognize digits written by hand.

Author: freeCodeCamp

Description: This course will teach you how to use Keras, a neural network API written in Python and integrated with TensorFlow. Students will learn how to prepare and process data for artificial neural networks, build and train artificial neural networks from scratch, build and train convolutional neural networks (CNNs), implement fine-tuning and transfer learning, and more.

Author: Edureka

Description: This Edureka deep learning full course video will help you understand and learn deep learning and TensorFlow are in detail. This deep learning tutorial is ideal for both beginners as well as professionals who want to master deep learning algorithms.

Author: Edureka

Description: This Edureka video will help you to understand the relationships between deep learning, machine learning, and artificial intelligence. This tutorial discusses AI, machine learning and its limitations, and how deep learning overcame machine learning limitations. Additional topics include deep learning applications and TensorFlow.

Author: freeCodeCamp

Description: Learn the fundamental concepts and terminology of seep learning, a sub-branch of machine learning. This course is designed for absolute beginners with no experience in programming. You will learn the key ideas behind deep learning without any code. It also covers neural networks and various machine learning constructs.

Tim is Solutions Review's Editorial Director and leads coverage on big data, business intelligence, and data analytics. A 2017 and 2018 Most Influential Business Journalist and 2021 "Who's Who" in data management and data integration, Tim is a recognized influencer and thought leader in enterprise business software. Reach him via tking at solutionsreview dot com.

Originally posted here:
The 6 Best Deep Learning Tutorials on YouTube to Watch Right Now - Solutions Review

Read More..

Agnostiq Announces Partnership With Mila to Bridge the Quantum Computing and Machine Learning Communities – PRNewswire

TORONTO, Jan. 18, 2022 /PRNewswire/ --Agnostiq, Inc., the first-of-its-kind quantum computing SaaS startup, is pleased to announce that it has formed a strategic partnership with Montreal-based Mila, the world's largest machine learning research institute, in an effort to bridge the gap between the quantum computing and machine learning communities.

"Quantum computing will have a tremendous impact on many fields and machine learning is no exception," says Oktay Goktas, CEO of Agnostiq. "A partnership with Mila brings us access to a world-class research community that comes with decades of experience in machine learning, which will in turn help us design better tools for emergent quantum machine learning use cases."

The new partnership gives Mila access to Agnostiq's quantum researchers, who are working on classes of machine learning problems that are specific to quantum computing, and Agnostiq access to Mila's AI/ML researchers and partner network. Partnering with Mila will help Agnostiq remain at the forefront and among the first to discover compelling new use cases for quantum machine learning.

"Agnostiq offers an exciting opportunity to explore ML challenges specific to quantum computing, as our strategic alliance with this promising startup will allow us to combine our expertise," says Stphane Ltourneau, Executive Vice President of Mila. "Mila's research community works daily toward improving the democratization of machine learning, developing new algorithms, and advancing deep learning capabilities. We are thrilled to work closely with Agnostiq to continue these important missions."

The partnership will also support Agnostiq's talent attraction and retention efforts, encouraging potential candidates to apply, as they will have the opportunity to collaborate with Mila's world-renowned researchers. Finally, the collaboration further validates Canada's position as a global leader in quantum computing and machine learning research.

ABOUT AGNOSTIQ INC.:Agnostiq develops software tools and applications for advanced computing resources, with the aim of making these technologies more accessible to developers and enterprise customers. Agnostiq is an interdisciplinary team of physicists, computer scientists, and mathematicians backed by leading investors in New York and Silicon Valley. Learn more at http://www.agnostiq.ai.

ABOUT MILA:Founded by Professor Yoshua Bengio of the Universit de Montral, Mila is a research institute in artificial intelligence which rallies about 700 researchers specializing in the field of deep learning. Based in Montreal, Mila's mission is to be a global pole for scientific advances that inspires innovation and the development of AI for the benefit of all. Mila is a non-profit organization recognized globally for its significant contributions to the field of deep learning, particularly in the areas of language modeling, machine translation, object recognition and generative models. Learn more at mila.quebec.

MEDIA CONTACT: Nina Pfister, MAG PR via email at [emailprotected]or by phone: 781-929-5620

SOURCE Agnostiq

Go here to read the rest:
Agnostiq Announces Partnership With Mila to Bridge the Quantum Computing and Machine Learning Communities - PRNewswire

Read More..

Machine learning assisted real-time deformability cytometry of CD34+ cells allows to identify patients with myelodysplastic syndromes | Scientific…

This paper provides a proof-of-concept to use RT-DC for detection of MDS. As RT-DC captures the morphology of cells, the information content is similar to morphology analyses of BM smears which is currently the gold standard for MDS diagnosis. In addition, the mechanical readout of RT-DC is a promising feature as earlier studies showed alterations of the actin cytoskeleton in association with MDS8,9,10.

Current MDS diagnosis routines are under reconsideration due to reproducibility issues, high labor intensiveness, and requirement of expert staff2,4,19. These issues could be addressed utilizing a combination of imaging flow cytometry (IFC) for high-throughput acquisition and machine learning for automated data analysis20,21. In IFC, fluorescence images are captured which allows labelling of different cell types and intracellular structures. However it was already shown that using deep learning, brightfield images are sufficient for example to predict lineages of differentiation or distinguish cell types in blood22,23. Hence, the label-free approach of RT-DC could be advantageous as the staining process can be omitted.

In the present work we employ RT-DC for the first time for detection of MDS. From each captured cell, seven features are computed in real-time, which were then used to train a random forest model, reaching an accuracy of 82.9% for the classification of healthy and MDS samples. As RT-DC performs image analysis in real-time, the MDS classification result could be provided immediately during the measurement. Both, the label-free aspect of RT-DC and the real-time analysis could allow to shorten the time needed for diagnosis.

By employing a model interpretation technique, we found that the width of the distribution of cell sizes is one of the most important criteria used by the random forest classification model. While employing only a single feature for classification lowers the accuracy (78%), it may be more suitable for observation in clinical practice. Interestingly, our finding is in accordance with the WHO guidelines which suggest a consideration of cell sizes during morphology evaluation. Our measurements show consistently that a subpopulation of cells in the size range (25 mu {text{m}}^{2}le Ale 45 mu {text{m}}^{2}) is underrepresented in MDS samples (see Fig.1D and Supplementary Fig. S3). This effect could be explained by the reduced number of B lymphocyte precursor cells in MDS24, which are CD34+ and could be present in the sample after CD34 based sorting25. Moreover, the histogram of cell sizes in Fig.1D shows a narrow peak at 50m2 for MDS, while the healthy counterpart presents a wider distribution. Hence, especially the width of the distribution plays a role, rather than the mean or median which is similar for both samples. However, since only 41 samples have been employed to train and validate the random forest model, the extrapolation of this study on the highly heterogeneous MDS population is limited, as the model could be overfitted to this small dataset. Moreover, random forest models do not perform well in extrapolation tasks. Hence, a larger prospective clinical study is required to reach more decisive conclusions.

Our work considered seven features obtained using RT-DC which can be summarized into three groups: features describing cell size (A, Lx, Ly), mechanical properties (, D, I), and porosity (). However, updated versions of the RT-DC technology are capable to save the brightfield image and compute transparency features in real-time which was shown to allow for discrimination between different blood cell types26. Moreover, images can be evaluated by a deep neural net which allows to employ fine grained details of the image for an accurate classification22,23. Future research should incorporate those new modalities to improve label-free detection of MDS using RT-DC.

MDS is caused by accumulation of genetic mutations which can be identified by whole genome sequencing. While costs for whole genome sequencing reduced from a hundred million to a thousand dollars during the last 20years, currently only targeted sequencing plays a role in clinical practice27. Here, only chosen genes that are frequently affected in MDS are checked, which is problematic, due to the large genomic heterogeneity present in various types of MDS28,29. Therefore, the standard diagnosis relies on an assessment of cell morphology as an indirect readout of genetic properties. Morphological alterations are accompanied by changes in the F-actin distribution and structural changes of the cytoskeleton8,9,10. RT-DC allows to measure mechanical properties of cells that are determined by the cytoskeleton5,30,31. It was already shown that diseases like malaria, leukemia, or spherocytosis lead to measurable differences in mechanical properties26,32. To link mechanical and genetic changes, we measured HSCs from MDS patients using RT-DC and performed molecular analysis in parallel. Figure2B indicates that larger numbers of genetic mutations correspond to lower median deformation. Therefore, RT-DC could provide an additional indirect readout of acquired mutations that has low cost per measurement, low measurement time, and offers real-time analysis results. However, despite the high correlation, we would regard this finding as hypothesis-generating due to the small sample size (n=10). Additionally, we could neither identify an association of mutation type and deformation, nor a significant mechanical difference between the low and the high-risk group (data not shown), but rather the biological features of the blast cells, such as number of mutations, correlated with the mechanical properties. The importance of Dmedian resulting from the random forest model is low (see Fig.1B). This suggests that Dmedian is similar for healthy and MDS samples. Hence, the approach of correlating Dmedian to infer the number of mutations is only valid for samples for which MDS had already been diagnosed.

HSCs only make up approximately 1% of the cells in the bone marrow33,34. To focus our study on this small subpopulation, we used MACS for CD34 enrichment of HSCs prior to the measurement. However, as the cells produced by mutated HSCs are presumably morphologically different from the healthy counterpart, a future endeavor should assess unsorted bone marrow in RT-DC using a similar approach as shown in the present work. Moreover, the efficiency of CD34 isolation is low, which results in small total numbers of cells for the measurement. As a result, our measurements could not fully employ the available throughput-capacity of RT-DC. Samples shown in this manuscript were subjected to cryopreservation and thawing which could potentially alter the cell morphology and MDS prediction outcome. A follow up project should therefore ideally use fresh BM.

Taken together, our study shows that RT-DC has the potential to expand the current status quo of MDS diagnostics. Both, morphological and mechanical readout from RT-DC are promising parameters for identification of MDS. Whether this method can be complementary to the standard diagnostic procedures in the borderline cases or serve as a rapid reliable test in the initial diagnostics remains to be demonstrated in the prospective clinical studies.

See the article here:
Machine learning assisted real-time deformability cytometry of CD34+ cells allows to identify patients with myelodysplastic syndromes | Scientific...

Read More..

6 Ways Machine Learning Can Improve Customer Satisfaction – TechSpective

Machine learning algorithms are assisting companies in improving and advancing many different aspects of their internal and external technological tools. One of the most common areas where machine learning is particularly useful is in customer service. Here are six ways machine learning can improve customer satisfaction.

These algorithms can more easily coordinate and handle workloads related to tools as varied as neural network chips, keyword searching algorithms, and data analytics programs. In general, machine learning can be leveraged to improve the efficiency and speed of these tools and all types of other processes and programs. With these algorithms, youre more likely to be able to quickly figure out what customers need and want and then direct them where they want to go.

The ability of machine learning algorithms to learn as they work allows them to more accurately and effectively pinpoint the needs of each customer and provide the necessary personalization and customization options. The algorithm can learn about each customer using your website and utilize the information it gleans to ensure the customer gets the experience or service that will most effectively benefit him or her. To take it a step further, as the algorithm learns more about customers and how to do its job efficiently, it may be capable of determining specific products that the customer is most likely to be interested in buying or using. It can effectively match users with the right products for them and their needs.

Not only can these algorithms improve your customers general experiences with your brand and with specific areas of your company, such as your website or your customer service department, but they can also help improve the customer experience in ways that are less immediately obvious. One such way is in improving your ability to identify fraud. These algorithms are capable of scanning and reviewing exponentially more transactions faster and more accurately than human beings can. Over time, your machine learning algorithm can learn to better identify signs of potential or definite fraud or identity theft.

One of the key uses of machine learning algorithms so far has been to perfect data collection, analytics and the generation of insights and predictions based on that data. You can use a machine learning algorithm to collect more cohesive data sets based on things your customers provide through data input, clicks or navigation around your website. The algorithm can then develop more complete insights and predictions based on that data in order to determine the best future marketing campaigns or customer service innovations and to identify valuable potential customers.

Another less obvious benefit of machine learning algorithms for your customers experiences is the enablement of continuous improvement these algorithms can provide. You can use the data collected by your algorithms to determine where your customer service is lacking and what you need to invest or do in order to make improvements. Powerful machine learning algorithms may even be capable of making certain kinds of adjustments and improvements automatically as they learn more about your customer service practices and your customers experiences with those practices and tools.

A powerful machine learning algorithm can even learn how to understand a customers intent when he or she interacts with your company on your website, via a phone call to your customer service department or over your social media. They can do so by gathering past user data related to the customer or by gathering information about the customers location regarding his or her situation or the product causing an issue. This is a common usage by energy companies, for example, when customers experience power outages.

Machine learning can help you improve your customer satisfaction results in various ways. These algorithms can provide you with the means to gather more data, interpret that data better, provide more personalization and continuously improve your offerings.

See the article here:
6 Ways Machine Learning Can Improve Customer Satisfaction - TechSpective

Read More..

Analyzing Twisted Graphene with Machine Learning and Raman Spectroscopy – AZoNano

In a recent study published in the journal ACS Applied Nano Materials,researchers developed a novel techniquefor automatically classifying the Raman spectra of twisted bilayer graphene (tBLG) into a variety of twist orientations using machine learning (ML).

Study:Machine Learning Determination of the Twist Angle of Bilayer Graphene by Raman Spectroscopy: Implications for van der Waals Heterostructures. Image Credit:ktsdesign/Shutterstock.com

With the recent surge in demand in tBLG, rapid, efficient, and non-destructive techniques for accurately determining the twist angle are needed.

Given the vast quantity of information on all aspects of the graphene recorded in its Raman spectrum, Raman spectroscopy might offer such a methodology. However, the angle determination can be difficult due to small variations in the Spectral data caused by the stacking sequence.

In recent decades, the discovery of highly conductive phases at tiny twist orientations has prompted interest in twisted bilayer graphene (tBLG) research. However, estimating the twist angle of tBLG layersisdifficult.

The most precise angle measurements are obtained using high-resolution imaging methods, such as transmission electron microscopy (TEM) or scanning probe microscopy (SPM).

The disadvantage is that observations take time and requireeither a free-standing material or one that is mounted on a current collector. Furthermore, they give highly localized features on sub-micron-sized locations, while the twist degree might change by several micrometers.

As a result, these approaches are inappropriate for real applications that need large-area categorizations on unpredictable materials in a short time.

Raman spectroscopy is a non-invasive examination method that allows for a variety of substrates and environments to be used for measurements, as well as the ability to analyze relatively vast regions in a small amount of time.

It has alsobeen extensively employed in graphene characterization, offering a wealth of data about the material's properties, purity, and electrical configuration.

In the case of tBLG, the twist orientation may also be determined using spectroscopic methods, which can provide sub-degree accuracy for specific angle regions.

In particular, calculating the twist angle necessitates an evaluation of numerous Raman spectra components at the same time. However, the growing complexity of the spectrum might make this process much more challenging.

This intricacy is most noticeable at low twist orientations when the tBLG experiences a structural rebuilding.

Although the Raman spectrum includes data on the angular position of tBLG, the differences for various angles can be every minute, including small alterations in the locations, breadth, and intensity ratios of the various peaks.

These distinctions are sometimes undetectable at first sight and may be easily ignored, necessitating a thorough examination of the spectrum.

Machine learning (ML) is a set of approaches that use mathematics to classify new information depending on a model's learning or recognize trends in uncategorized data (unsupervised ML).

ML-based approaches are being prepared and implemented in many parts of 2D material study and processing. ML has lately been shown to be useful in identifying the twist angle of simulated Raman spectrum parts. ML was also utilized to detect certain twist angles of BLG created by synthetic single-layer graphene stacks.

To estimate the layering sequence of tBLG from its absorption spectra, the researchers created a simple, rapid, and low computationally intensive ML-based analytical technique in this study. This approach entails gathering enough data from the tBLG Raman spectra to build an ML model capable of inferring the twist angle within prescribed ranges.

Compared to manual twist angle labeling, the suggested approach achieves a precision of over 99 percent. Furthermore, the approach is computationally light, delivering forecasts for whole Raman mapping, including dozens of wavelengths in a matter of a few seconds on even the most basic desktop machines.

Finally, the suggested method's versatility allows it to be expanded to measure the amount of strain and loading in graphene and the adjustment factor of other thin films and heterostacks.

The suggested approach might also be used to investigate the twist orientations of various vdW nanostructured materials, making it a valuable and straightforward analytical tool with real-world applications for the current level of understanding of twistronics.

Continue reading: How Graphene was Found in 4.5 Billion-Year-Old Meteorites.

Sols-Fernndez, P. et al. (2022). Machine Learning Determination of the Twist Angle of Bilayer Graphene by Raman Spectroscopy: Implications for van der Waals Heterostructures. ACS Applied Nano Materials. Available at: https://pubs.acs.org/doi/10.1021/acsanm.1c03928

Disclaimer: The views expressed here are those of the author expressed in their private capacity and do not necessarily represent the views of AZoM.com Limited T/A AZoNetwork the owner and operator of this website. This disclaimer forms part of the Terms and conditions of use of this website.

More here:
Analyzing Twisted Graphene with Machine Learning and Raman Spectroscopy - AZoNano

Read More..

Cloudian Partners with WEKA to Deliver High-Performance, Exabyte-Scalable Storage for AI, Machine Learning and Other Advanced Analytics -…

SAN MATEO, Calif., Jan. 20, 2022 (GLOBE NEWSWIRE) -- Cloudian today announced the integration of its HyperStore object storage with the WEKA Data Platform for AI, providing high-performance, exabyte-scalable private cloud storage for processing iterative analytical workloads. The combined solution unifies and simplifies the data pipeline for performance-intensive workloads and accelerated DataOps, all easily managed under a single namespace. In addition, the new solution reduces the storage TCO associated with data analytics by a third, compared to traditional storage systems.

Advanced Analytics Workloads Create Data Storage ChallengesOrganizations are consuming and creating more data than ever before, and many are applying AI, machine learning (ML) and other advanced analytics on these large data sets to make better decisions in real-time and unlock new revenue streams. These analytics workloads create and use massive data sets that pose significant storage challenges, most importantly the ability to manage the data growth and enable users to extract timely insights from that data. Traditional storage systems simply cant handle the processing needs or the scalability required for iterative analytics workloads and introduce bottlenecks to productivity and data-driven decision making.

Cloudian-WEKA Next Generation Storage PlatformTogether, Cloudian and WEKA enable organizations to overcome the challenges of accelerating and scaling their data pipelines while lowering data analytics storage costs. WEKAs data platform, built on WekaFS, addresses the storage challenges posed by todays enterprise AI workloads and other high-performance applications running on-premises, in the cloud or bursting between platforms. The joint solution offers the simplicity of NAS, the performance of SAN or DAS and the scale of object storage, along with accelerating every stage of the data pipeline from data ingestion to cleansing to modeled results.

Integrated through WEKAs tiering function, Cloudians enterprise-grade, software-defined object storage provides the following key benefits:

As organizations increasingly employ AI, ML and other advanced analytics to extract greater value from their data, they need a modern storage platform that enables fast, easy data processing and management, said Jonathan Martin, president, WEKA. The combination of the WEKA Data Platform and Cloudian object storage provides an ideal solution that can seamlessly and cost-effectively scale to meet growing demands.

When it comes to supporting advanced analytics applications, users shouldnt have to make tradeoffs between storage performance and capacity, said Jon Toor, chief marketing officer, Cloudian. By eliminating any need to compromise, the integration of our HyperStore software with the WEKA Data Platform gives customers a storage foundation that enables them to fully leverage these applications so they can gain new insights from their data and drive greater business and operational success.

The new solution is available today. For more information, visit cloudian.com/weka/.

About CloudianCloudian is the most widely deployed independent provider of object storage. With a native S3 API, it brings the scalability and flexibility of public cloud storage into the data center while providing ransomware protection and reducing TCO by up to 70% compared to traditional SAN/NAS and public cloud. The geo-distributed architecture enables users to manage and protect object and file data across siteson-premises and in the cloudfrom a single platform. Available as software or appliances, Cloudian supports conventional and containerized applications. More atcloudian.com.

U.S. Media ContactJordan Tewell 10Fold Communicationscloudian@10fold.com+1 415-666-6066

EMEA Media Contact Jacob GreenwoodRed Lorry Yellow Lorrycloudian@rlyl.com+44 (0) 20 7403 8878

See original here:
Cloudian Partners with WEKA to Deliver High-Performance, Exabyte-Scalable Storage for AI, Machine Learning and Other Advanced Analytics -...

Read More..