Category Archives: Machine Learning

How quantum computing is helping businesses to meet objectives – Information Age

Johannes Oberreuter, Quantum Computing practice lead and data scientist at Reply, spoke to Information Age about how quantum computing is helping businesses to meet objectives

Quantum is emerging as a new vehicle for business problem solving.

Quantum computing is an evolving technology that promises to enhance an array of business operations. Based on quantum mechanics that focus on the smallest dimensions of nature molecules, atoms and subatomic particles quantum computers are set to provide faster solutions to complex business problems, through testing multiple possible solutions for a problem simultaneously.

The basis for quantum computing is a unit of information known as a qubit; unlike bits, which can only have the values zero or one, can come in the form of anything in between, which allows for this new approach to become possible, and is called a superposition. Combined, multiple qubits can produce many outcomes at the same time. Every extra qubit doubles the search space, which therefore grows exponentially.

Many companies are looking into how quantum can bolster industries and provide new use cases for businesses. One organisation thats exploring this space is Reply, which has been developing solutions for optimisation in logistics, portfolio management and fault detection, among other areas.

Discussing how Reply is helping to provide possible use cases to its clients, quantum computing expert Johannes Oberreuter said: We work on a level which translates the problem into a quantum language that is as universal as possible, and doesnt go too deep into the hardware.

The first thing weve found thats delivering value now is the domain of optimisation problems. An example is the travelling salesman problem, which has lots of applications in logistics, where complexities and constraints also need to be accounted for, like during the pandemic.

Very often, problems, which are found too complex to be optimised on common hardware, are tackled by some heuristics. Usually, theres a team or a person with experience in the domain, who can help with this, but they dont know yet that there are better solutions out there now. Quantum computing allows for problems being presented in a structured way similar to a wish list, containing all business complexities. They are all encoded into a so-called objective function, which can then be solved in a structured way.

Companies have used all sorts of algorithms and brain power to try to solve optimisation problems. Finding the optimum with an objective function is still a difficult problem to solve, but here a quantum computer can come to the rescue.

Pushing parameters

According to Oberreuter, once a quantum computer becomes involved in the problem solving process, the optimal solution can really be found, allowing businesses to find the best arrangements for the problem. While current quantum computers, which are suitable for this kind of problems, called quantum annealers now have over 5,000 qubits, many companies that enlist Replys services often find that problems they have require more than 16,000-20,000 variables, which calls for more progress to be made in the space.

You can solve this by making approximations, commented the Reply data scientist. Weve been writing a program that is determining an approximate solution of this objective function, and we have tested it beyond the usual number of qubits needed.

The system is set up in a way that prevents running time from increasing exponentially, which results in a business-friendly running time of a couple of seconds. This reduces the quality of the solution, but we get a 10-15% better result than what business heuristics are typically providing.

Through proofs-of-concepts, Reply has been able to help clients to overcome the challenge of a lack of expertise in quantum. By utilising and building up experience in the field, a shoulder-to-shoulder approach helps to clarify how solutions can be developed more efficiently.

Machine learning has risen in prominence over the last few years to aid automation of business processes with data, and help organisations meet goals faster. However, machine learning projects can sometimes suffer from lack of data and computational expense. To combat this, Reply has been looking to the problem solving capabilities brought by quantum computing.

Oberreuter explained: What weve discovered with quantum machine learning is you can find better solutions, even with the limited hardware thats accessible currently. While there will probably never be an end-to-end quantum machine learning workflow, integration of quantum computing into the current machine learning workflow is useful.

Some cloud vendors now offer quantum processing units (QPUs). In a deep learning setup for complex tasks, you could easily rent it from the cloud providers by individual calls to experiment, if it improves your current model.

What weve found interesting from our contribution towards the quantum challenge undertaken by BMW and AWS, is the marriage of classical machine learning models with quantum models. The former is really good at extracting attributes from unstructured data such as images, which are then joined by a quantum representation which provides an advantage for classification.

How organisations can drive value from AI on the edge

Mike Ellerton, partner at Go Reply, spoke to Information Age about Replys recent research conducted into edge AI, and how organisations can drive value from the technology. Read here

Additionally, quantum technologies are being explored for cyber security, with the view that soon quantum computers can solve problems that are currently insurmountable for todays technologies. A particular algorithm thats been cited by Reply, that could be solved by quantum computing, is the one used for RSA key cryptography, which while trusted to be secure now, is estimated to need 6000 error-free qubits to be cracked in the space of two weeks.

Quantum technology for cyber security is now on the shelf, and were offering this to our clients to defend against this threat, said Oberreuter. Quantum mechanics have a so-called no-cloning theorem, which prevents users from copying messages sent across a communication channel. The crux is that in order for this to work, you need a specialised quantum channel.

We have experts who specialise in cyber security, that have been leading the effort to craft an offering for this.

Reply is a network of highly specialised industry companies, that helps clients across an array of sectors to optimise and integrate processes, applications and devices using the latest technologies. Established in 1996, the organisation offers services for capabilities including quantum, artificial intelligence (AI), big data, cloud and the Internet of Things (IoT). More information on the services that Reply provides can be found here.

This article was written as part of a paid-for content campaign with Reply

View post:
How quantum computing is helping businesses to meet objectives - Information Age

A machine learning model based on tumor and immune biomarkers to predict undetectable MRD and survival outcomes in multiple myeloma – DocWire News

This article was originally published here

Clin Cancer Res. 2022 Jan 21:clincanres.3430.2021. doi: 10.1158/1078-0432.CCR-21-3430. Online ahead of print.

ABSTRACT

PURPOSE: Undetectable measurable residual disease (MRD) is a surrogate of prolonged survival in multiple myeloma (MM). Thus, treatment individualization based on the probability of a patient to achieve undetectable MRD with a singular regimen, could represent a new concept towards personalized treatment with fast assessment of its success. This has never been investigated; therefore, we sought to define a machine learning model to predict undetectable MRD at the onset of MM.

EXPERIMENTAL DESIGN: This study included 487 newly-diagnosed MM patients. The training (n=152) and internal validation cohort (n=149) consisted of 301 transplant-eligible active MM patients enrolled in the GEM2012MENOS65 trial. Two external validation cohorts were defined by 76 high-risk transplant-eligible smoldering MM patients enrolled in the GEM-CESAR trial, and 110 transplant-ineligible elderly patients enrolled in the GEM-CLARIDEX trial.

RESULTS: The most effective model to predict MRD status resulted from integrating cytogenetic [t(4;14) and/or del(17p13)], tumor burden (bone marrow plasma cell clonality and circulating tumor cells) and immune-related biomarkers. Accurate predictions of MRD outcomes were achieved in 71% of cases in the GEM2012MENOS65 trial (n=214/301), and 72% in the external validation cohorts (n=134/186). The model also predicted sustained MRD negativity from consolidation onto 2-years maintenance (GEM2014MAIN). High-confidence prediction of undetectable MRD at diagnosis identified a subgroup of active MM patients with 80% and 93% progression-free and overall survival rates at five years.

CONCLUSION: It is possible to accurately predict MRD outcomes using an integrative, weighted model defined by machine learning algorithms. This is a new concept towards individualized treatment in MM.

PMID:35063966 | DOI:10.1158/1078-0432.CCR-21-3430

Link:
A machine learning model based on tumor and immune biomarkers to predict undetectable MRD and survival outcomes in multiple myeloma - DocWire News

Associate / Full Professor of Theoretical Biophysics and Machine Learning job with RADBOUD UNIVERSITY NIJMEGEN | 278686 – Times Higher Education (THE)

Associate / Full Professor of Theoretical Biophysics and Machine Learning

A world from which we demand more and more requires people who can make a contribution. Critical thinkers who will take a closer look at what is really important. As a Professor, you will perform leading research and teach students in the area of theoretical biophysics and physics-based machine learning, to strengthen the role and visibility of the international Theoretical Biophysics landscape.

As a successful candidate you will join the Department of Biophysics at the Donders Center for Neuroscience (DCN) and perform internationally leading theoretical research in an area of theoretical biophysics or physics-based machine learning. You are interested in applications of theoretical biophysics methods to neuroscience problems studied in the DCN, and you will engage actively in interdisciplinary research collaborations with other physicists in the Faculty of Science and with external partners. You will contribute to the teaching and the innovation of Radboud's popular theoretical machine learning and biophysics courses, and possibly contribute to other core undergraduate physics subjects taught at the Faculty of Science. You will supervise students' research projects at the Bachelor's, Master's and PhD levels. Finally, you will contribute to the effective administration of Radboud University and the acquisition of research funding, and will strengthen the role and visibility of Radboud University in the international Theoretical Biophysics landscape.

Profile

We are

The Donders Institute for Brain, Cognition and Behaviour of Radboud University seeks to appoint a Professor of Theoretical Biophysics and Machine Learning. The Donders Institute is a world-class research institute, housing more than 700 researchers devoted to understanding the mechanistic underpinnings of the human mind/brain. Research at the Donders Institute focuses on four themes:

Language and Communication

Perception, Action, and Decision-making

Development and Lifelong Plasticity

Natural Computing and Neurotechnology.

We have excellent and state-of-the-art research facilities available for a broad range of neuroscience research. The Donders Institute fosters a collaborative, multidisciplinary, supportive research environment with a diverse international staff. English is the lingua franca at the Institute.

You will join the academic staff of the Donders Center for Neuroscience (DCN) - one of the four Donders Centers at Radboud University's Faculty of Science. The Biophysics Department is part of the DCN. Neurophysicists at DCN mainly conduct experimental, theoretical and computational research into the principles of information processing by the brain, with particular focus on the mammalian auditory and visual systems. The Physics of Machine Learning and Complex Systems Group studies a broad range of theoretical topics, ranging from physics-based machine learning paradigms and quantum machine learning, via Bayesian inference and applications of statistical mechanics techniques in medical statistics, to network theory and the modelling of heterogeneous many-variable processes in physics and biology. The group engages in multiple national and international research collaborations, and participates in several multidisciplinary initiatives that support theoretical biophysics and machine learning research and teaching at Radboud University.

Radboud University actively supports equality, diversity and inclusion, and encourages applications from all sections of society. The university offers customised facilities to better align work and private life. Parents are entitled to partly paid parental leave and Radboud University employees enjoy flexibility in the way they structure their work. The university highly values the career development of its staff, which is facilitated by a variety of programmes. The Faculty of Science is an equal opportunity employer, committed to building a culturally diverse intellectual community, and as such encourages applications from women and minorities.

Radboud University

We want to get the best out of science, others and ourselves. Why? Because this is what the world around us desperately needs. Leading research and education make an indispensable contribution to a healthy, free world with equal opportunities for all. This is what unites the more than 24,000 students and 5,600 employees at Radboud University. And this requires even more talent, collaboration and lifelong learning. You have a part to play!

We offer

Additional employment conditions

Work and science require good employment practices. This is reflected in Radboud University's primary and secondary employment conditions. You can make arrangements for the best possible work-life balance with flexible working hours, various leave arrangements and working from home. You are also able to compose part of your employment conditions yourself, for example, exchange income for extra leave days and receive a reimbursement for your sports subscription. And of course, we offer a good pension plan. You are given plenty of room and responsibility to develop your talents and realise your ambitions. Therefore, we provide various training and development schemes.

Would you like more information?

For questions about the position, please contact Ton Coolen, Professor at +31 24 361 42 45 or ton.coolen@donders.ru.nl.

Practical information and applications

You can apply until 25 February 2022, exclusively using the button below. Kindly address your application to Ton Coolen. Please fill in the application form and attach the following documents:

The first round of interviews will take place around the end of March. You would preferably begin employment on 1 September 2022.

This vacancy was also published in a slightly modified form in 2021. Applicants who were rejected at that time are kindly requested not to apply again.

We can imagine you're curious about our application procedure. It offers a rough outline of what you can expect during the application process, how we handle your personal data and how we deal with internal and external candidates.

We drafted this vacancy to find and hire our new colleague ourselves. Recruitment agencies are kindly requested to refrain from responding.

The rest is here:
Associate / Full Professor of Theoretical Biophysics and Machine Learning job with RADBOUD UNIVERSITY NIJMEGEN | 278686 - Times Higher Education (THE)

Heard on the Street 1/24/2022 – insideBIGDATA

Welcome to insideBIGDATAs Heard on the Street round-up column! In this regular feature, we highlight thought-leadership commentaries from members of the big data ecosystem. Each edition covers the trends of the day with compelling perspectives that can provide important insights to give you a competitive advantage in the marketplace. We invite submissions with a focus on our favored technology topics areas: big data, data science, machine learning, AI and deep learning. Enjoy!

COVID-19: A Data Tsunami That Ushered in Unprecedented Opportunities for Businesses and Data Scientists. Commentary by Thomas Hazel, founder & CTO at ChaosSearch

From creating volatile data resources to negatively impacting forecasting models, there have been countless challenges the pandemic has caused for organizations that rely on data to inform business decisions. However, there is also an upside to the data tsunami that COVID-19 created. The movement to all-things-digital translated into a tsunami of log data streaming from these digital systems. All this data presented an incredible opportunity for companies to deeply understand their customers and then tailor customer and product experiences. However, theyd need the right tools and processes in place to avoid being overwhelmed by the volume of data. The impact spans all industries, from retail to insurance to education.Blackboard is a perfect example. The world-leading EdTech provider was initially challenged at the start of the pandemic with the surge of daily log volumes from students and school systems that moved online seemingly overnight. The company quickly realized they needed a way to efficiently analyze log data for real-time alerts and troubleshooting, as well as a method to access long-term data for compliance purposes. To accomplish this, Blackboard leverages its data lake to monitor cloud deployments, troubleshoot application issues, maximize uptime, and deliver on data integrity and governance for highly sensitive education data. This use case demonstrates just how important data has become to organizations that rely on digital infrastructure and how a strong data platform is a must to reduce the time, cost, and complexity of extracting insights from data. While the pandemic created this initial data tsunami, tech-driven organizations that have evolved to capitalize on its benefits, like Blackboard, have accepted that this wave of data is now a constant force that they will have to manage more effectively for the foreseeable future.

Cloud Tagging Best Practices. Commentary by Keith Neilson, Technical Evangelist at CloudSphere

While digital transformation has been on many organizations priority list for years, the Covid-19 pandemic applied more pressure and urgency to move this forward. Through their modernization efforts, companies have unfortunately wasted time and resources on unsuccessful data deployments, ultimately jeopardizing company security. For optimal cyber asset management, consider the following cloud tagging best practices:Take an algorithmic approach to tagging. While tags can represent simple attributes of an asset (like region, department, or owner), they can also assign policies to the asset. This way, assets can be effectively governed, even on a dynamic and elastic platform. Next, optimize tagging for automation and scalability. Proper tagging will allow for vigorous infrastructure provisioning for IT financial management, greater scalability and automated reporting for better security. Finally, be sure to implement consistent cloud tagging processes and parameters within your organization. Designate a representative to enforce certain tagging formulas, retroactively tag when IT personnel may have added assets or functions that they didnt think to tag and reevaluate business outputs to ensure tags are effective.While many underestimate just how powerful cloud tagging can be, the companies embracing this practice will ultimately experience better data organization, security, governance and system performance.

Using AI to improve the supply chain.Commentary by Melisa Tokmak, GM of Document AI, Scale AI

As supply chain delays continue to threaten businesses at the beginning of 2022, AI can be a crucial tool for logistics companies to speed up their supply chain as the pandemic persists. Logistics and freight forwarding companies are required to process dozens of documents such as bills of lading, commercial invoices and arrival notices fast, and with the utmost accuracy, in order to report data to Customs, understand changing delivery timelines, collect & analyze data about moving goods to paint information about the global trade. For already overtaxed and paperwork-heavy systems, manual processing and human error are some of the most common points of failure, which exacerbate shipping delays and result in late cargo, delayed cash flow & hefty fines.As logistics companies have a wealth of information buried in the documents they process, updating databases with this information is necessary to make supply chains more predictable globally. Most companies spend valuable time analyzing inconsistent data or navigating OCR and template-based solutions, which arent effective due to the high variability of data in these documents. Machine learning-based, end-to-end document processing solutions, such as Scale AIs Document AI, dont rely on templates and can automate this process; AI solutions allow logistics companies to leverage the latest industry research without changing their developer environment. This way, companies can focus on using their data to cater to customers and serve the entire logistics industry, rather than spending valuable time and resources on data-mining.ML-based solutions can extract the most valuable information accurately in seconds, accelerating internal operations, reducing the number of times containers are opened for checksdecreasing costs and shipping delays significantly. Using Scales Document AI, freight forwarding leader Flexport achieved significant cost savings in operations and decreased the processing time of each document. Flexports documents were formerly processed in over two days, but with Document AI, were processed in less than 60 seconds with 95%+ accuracy, all without having to build and maintain a team of machine learning engineers and data scientists. As COVID has led to a breakdown of internal processes, AI-powered document processing solutions are helping build systems back up: optimizing operations to handle any logistic needs that come their way at such a crucial time.

IBM to Sell Watson Health. Paddy Padmanabhan, Founder and CEO of Damo Consulting

IBMs decision to sell the Watson Health assets is not an indictment of the promise of AI in healthcare. Our research indicates AI was one of the top technology investments for health systems in 2021. Sure, there are challenges such as data quality and bias in the application of AI in the healthcare context but by and large there has been progress with AI in healthcare. The emergence of other players, notably Google with its Mayo Partnership, or Microsoft with its partnership with healthcare industry consortium Truveta are strong indicators of progress.

Data Privacy Day 2022 Commentary. Commentary by Lewis Carr, Senior Director, Product Marketing at Actian

In 2022, expect to see all personal information and data sharing options get more granular as to how we control them both on our devices and in the cloud specific to each company, school or government agency. Well also start to get some visibility into and control over how our data is shared between organizations without us involved. Companies and public sector organizations will begin to pivot away from the binary options (opt-in or opt-out) tied to a lengthy legal letter that no one will read and will instead provide the data management and cybersecurity platforms with granular permission to parts of your personal data, such as where its stored, for how long, and under what circumstances it can be used. You can also expect new service companies to sprout up that will offer intermediary support to monitor and manage your data privacy across.

Data Privacy Day 2022 Commentary. Commentary by Rob Price, Principal Expert Solution Consultant at Snow Software

The adoption of cloud technology has been a critical component to how we approach privacy and data protection today. A common misconception is that if your data is offsite or cloud-based its not your problem but that is not true because the cloud is not a data management system. Two fundamental factors for data protection and security are the recovery point objective (how old can data be when you recover it) and the recovery time objective (how quickly can you recover the data). Every companys needs are different, but these two factors are important when planning for data loss.

Sign up for the free insideBIGDATAnewsletter.

Join us on Twitter:@InsideBigData1 https://twitter.com/InsideBigData1

Read more here:
Heard on the Street 1/24/2022 - insideBIGDATA

Collaboration with NTT Research to advance computational neurobiology – Harvard Office of Technology Development

January 24, 2022 - Neurobiologists at Harvard University have entered a joint research agreement with NTT Research, Inc., a division of NTT, to study animal neuro-responses with the hope of informing future artificial intelligence systems. The five-year research project, launched in the fall of 2021, enables researchers at the two organizations to collaboratively study how animals maintain behavioral flexibility, specifically in the task of navigation. Greater understanding of how this challenge is approached in biology may eventually enable the design of new computing machines with similar capabilities. The agreement was coordinated by Harvard Office of Technology Development.

The principal investigator is Venkatesh Murthy, PhD, the Raymond Leo Erikson Life Sciences Professor of Molecular and Cellular Biology at Harvard and the Paul J. Finnegan Family Director of the Center for Brain Science. Murthys counterpart at NTT Research for the joint project is Physics & Informatics (PHI) Lab Research Scientist Gautam Reddy, PhD, who was previously an Independent Post-Doctoral Fellow at Harvards NSF-Simons Center for Mathematical and Statistical Analysis of Biology.

This joint research aims to better elucidate how animals maintain the ability to respond appropriately to a wide variety of complex real-world scenarios. The investigators expect the results from one aspect of the research to be a source of new, biologically inspired ideas for artificial reinforcement learning systems that rely on representation learning. Such ideas have played a major role in recent advances in artificial intelligence. Results from another aspect of the research should provide a quantitative understanding of how animals track trails, as well as identify the basic elements of general behavioral strategies that perform flexibly and reliably in the real world. Murthys lab has a long track record in experimental and computational neurobiology. Expertise relevant to the joint research includes the ability to record from or image many individual neurons in the brain while an animal performs behavioral tasks. This technical expertise will enable the research team to understand what computations are performed by biological neural networks when an animal is navigating in a complex world.

Murthy and Reddy have previously worked together on understanding the computational principles behind olfaction. Their focus was on how the smell receptors in the nose respond to blends of odorous compounds. During his time at Harvards NSF-Simons Center for Mathematical Biology, Reddy worked on the theory behind how animals track scent trails and on developing a computational framework to explain how evolution optimizes organisms.

I am delighted to continue this line of inquiry with Dr. Reddy through the NTT Research PHI Lab, Murthy said. The brain is an example of an extremely efficient computational device, and plenty of phenomena within it remain unexplored and unexplained. We believe the results of these investigations in neurobiology will reveal basic understandings and prove useful in the field of artificial intelligence.

Efficient computation is at the heart of quantum computing and neuroscience. Inspired by neuroscience, recent advances in machine learning have recently begun to change how we process data, said NTTs PHI Lab Director Yoshihisa Yamamoto, PhD. This joint research project could provide a rich source of animal-inspired algorithms that generalize across various research domains within NTT and inspire truly novel interdisciplinary ideas.

Adapted from a press release by NTT Research.

Here is the original post:
Collaboration with NTT Research to advance computational neurobiology - Harvard Office of Technology Development

Dask-ML dask-ml 2021.11.31 documentation

Dask-ML provides scalable machine learning in Python using Dask alongsidepopular machine learning libraries like Scikit-Learn, XGBoost, and others.

People may run into scaling challenges along a couple dimensions, and Dask-MLoffers tools for addressing each.

The first kind of scaling challenge comes when from your models growing solarge or complex that it affects your workflow (shown along the vertical axisabove). Under this scaling challenge tasks like model training, prediction, orevaluation steps will (eventually) complete, they just take too long. Youvebecome compute bound.

To address these challenges youd continue to use the collections you know andlove (like the NumPy ndarray, pandas DataFrame, or XGBoost DMatrix)and use a Dask Cluster to parallelize the workload on many machines. Theparallelization can occur through one of our integrations (like Dasksjoblib backend to parallelize Scikit-Learn directly) or one ofDask-MLs estimators (like our hyper-parameter optimizers).

The second type of scaling challenge people face is when their datasets growlarger than RAM (shown along the horizontal axis above). Under this scalingchallenge, even loading the data into NumPy or pandas becomes impossible.

To address these challenges, youd use Dasks one of Dasks high-levelcollections like(Dask Array, Dask DataFrame or Dask Bag) combined with one of Dask-MLsestimators that are designed to work with Dask collections. For example youmight use Dask Array and one of our preprocessing estimators indask_ml.preprocessing, or one of our ensemble methods indask_ml.ensemble.

Its worth emphasizing that not everyone needs scalable machine learning. Toolslike sampling can be effective. Always plot your learning curve.

In all cases Dask-ML endeavors to provide a single unified interface around thefamiliar NumPy, Pandas, and Scikit-Learn APIs. Users familiar withScikit-Learn should feel at home with Dask-ML.

Other machine learning libraries like XGBoost already havedistributed solutions that work quite well. Dask-ML makes no attempt tore-implement these systems. Instead, Dask-ML makes it easy to use normal Daskworkflows to prepare and set up data, then it deploys XGBoostalongside Dask, and hands the data over.

See Dask-ML + XGBoost for more information.

View original post here:
Dask-ML dask-ml 2021.11.31 documentation

How Snapchat Is Using AI And Machine Learning To Thwart Drug Deals – Hot Hardware

Snapchat is taking a proactive approach in fighting drug deals taking place on its social media platform. The company shared an update concerning its most recent efforts to halt the push to sell drugs through connections on its app.Snapchat typically makes the news when the social media platform goes dark, sending users in a frenzy wondering when their beloved app will be back up and running. But the extremely popular app, especially among teenagers and young adults, found itself in a different type of spotlight last October when NBC News did a story about the troubling drug deals presumably taking place on the app.

The report examined the death of teens and young adults who were suspected of buying fentanyl-laced drugs using Snapchat. In the report, it spoke about teens and young adults who had bought what they believed to be a prescription pill, but turned out to be a counterfeit pill containing deadly doses of fentanyl. Since that report, Snapchat has been ramping up its efforts to thwart drug deals on its platform.

Snap stated that it has a zero tolerance for drug dealing on Snapchat. It says it has made significant operational improvements over the past year toward its goal of completely eradicating drug dealers from its platform. It claims to take a holistic approach, which includes "deploying tools that proactively detect drug-related content, working with law enforcement to support their investigations, and provide in-app information and support to Snapchatters who search for drug-related terms through a new educational portal, Heads Up."

The social media company announced that it is adding two new partners to its Heads Up portal in order to provide important in-app resources to its users. Community Anti-Drug Coalitions of America (CADCA), is a nonprofit organization that focuses on creating safe, healthy and drug-free communities. Truth Initiative is the second addition, and is an organization that strives to steer teens and young adults away from smoking, vaping and nicotine in general. Along with these two new organizations being added, Snap will soon be releasing its next episode of it special Good Luck America series which will focus on fentanyl.

Snapchat is also updating its Quick Add suggestion feature in order to reduce interactions between kids and strangers. The company states, "In order to be discoverable in Quick Add by someone else, users under 18 will need to have a certain number of friends in common with that person." In the past, users would be given a list of recommended friends based on mutual connections, regardless if you knew the person in real life or not. Work is also being done on additional parental tools that it will roll out in the coming months.

Another way Snapchat is looking to deter drug dealers from using its platform, is in its cooperation with law enforcement. It has implemented measures using artificial intelligence (AI) and machine learning to identify drug slang and content on the app, and then works with law enforcement to report potential cases and to comply with information requests. Snapchat has increased its law enforcement operations team by 74% since its creation. And remarkably, Snapchat claims that a whopping 88% of drug related content it uncovers is proactively detected by its AI and machine learning algorithms. That's up from 33% since its previous update.

"When we find drug dealing activity, we promptly ban the account, use technology to block the offender from creating new accounts on Snapchat, and in some cases proactively refer the account to law enforcement for investigation," Snapchat says.

Continued here:
How Snapchat Is Using AI And Machine Learning To Thwart Drug Deals - Hot Hardware

Data Vault Holdings Expands Expertise In Artificial Intelligence, Machine Learning, and Big Data; Appoints Tony Evans of C3 AI To Advisory Board -…

NEW YORK, Jan. 19, 2022 /PRNewswire/ --Data Vault Holdings Inc., the emerging leader in metaverse data visualization, valuation, and monetization announced today the appointment of Tony Evans, General Manager of Financial Services for C3 AI (NYSE: AI), to its advisory board, fortifying Data Vault Holding's expertise in artificial intelligence, machine learning, fintech, e-commerce and security. A preeminent expert in business and sales, Mr. Evans has developed and executed transformative, customer-focused strategies across industries. From artificial intelligence to cybersecurity to e-commerce, he has managed global sales and partnership development, led global banking teams, driven growth, and developed customer big data and innovation strategies. As a member of the advisory board, Mr. Evans will advise Datavaultleadership on the automation and scale of their comprehensive crypto data solution.

"In my role at C3 AI, I witness daily the power data assets and tokenomics can play in the foundation for predictive technology that influences decisions and leads to disruption of incumbent markets. Data has now become both an indicator of business intelligence and a form of capital, and we can use this information to inform business innovation. Datavaultexpertly combines artificial intelligence, machine learning, and crypto-technology to transform data into salable business growth and revenues. I am honored to provide Datavault's leadership with perspective on emerging trends, market impact, and consumer issues in payments, AI, and data," says Tony Evans, General Manager of Financial Services for C3 AI.

As General Manager of Financial Services of leading enterprise AI software provider C3 AI, Mr. Evans directs financial services strategy, global sales, and partnership development. His expertise supports the delivery of the cross-industries enterprise platform C3 AI Suite, which enables businesses to develop, deploy, and operate large-scale AI, predictive analytics, and Internet of Things (IoT) applications. Mr. Evans' diversified background in the financial and technology sectors skillfully positions him to provide counsel to the executive team of Data Vault Holdings, as they develop and launch new products, design new revenue models, and simplify data visualization, valuation, and monetization processes layering effects through automation of their novel crypto-technologies. Additionally, Mr. Evans has served as Leader of Global Banking and Payments and Head of Financial Services (UK) with Amazon Web Services (AWS) (NASDAQ: AMZN); Head of Leonardo and Analytics (UK and Ireland) and SVP & Chief Operating Officer of Financial Services for SAP; and Managing Director (US) of BlackBerry. He has also served in leadership roles with Datawatch Corporation, Oracle, Applied Knowledge LTD, Visusol Consulting, and Smith Industries.

Mr. Evans holds an MBA specializing in business growth, change movement, and change strategy from the University of Brighton.

In coming weeks, Data Vault Holdings plans to announce additional members of its advisory board, with Ed Cushing, Global Account Manager at Amazon Web Services (AWS),recently announced as its inaugural member. New advisory board members will further aid in providing market insights, analytics expertise, and business and data monetization strategies through the use of Datavault's patented, cloud-based SaaS platform.

About Data Vault Holdings Inc.

Data Vault Holdings Inc. is a technology holding company that provides a proprietary, cloud-based platform for the delivery of branded data-backed cryptocurrencies. Data Vault Holdings Inc. provides businesses with the tools to monetize data assets securely over its Information Data Exchange(IDE). The company is in the process of finalizing the consolidation of its affiliates Data Donate Technologies, Inc., ADIO LLC, and Datavault Inc. as wholly owned subsidiaries under one corporate structure. Learn more about Data Vault Holdings Inc. here.

SOURCE Data Vault Holdings Inc.

Originally posted here:
Data Vault Holdings Expands Expertise In Artificial Intelligence, Machine Learning, and Big Data; Appoints Tony Evans of C3 AI To Advisory Board -...

Application of ensemble machine learning algorithms on lifestyle factors and wearables for cardiovascular risk prediction | Scientific Reports -…

Data source and study population

Data used in this study was drawn from a SingHEART prospective longitudinal cohort study (ClinicalTrials.gov Identifier: NCT02791152). The study is a multi-ethnic population-based study conducted on healthy Asians, aged 2169years old without known diabetes mellitus or prior cardiovascular disease (Ischemic heart disease, stroke, peripheral vascular disease). The study complied with the Declaration of Helsinki and written informed consent were given by participants. The study was approved by the SingHealth Centralized Institutional Review Board.

We included 600 volunteers, aged of 30years with valid calcium score, into the main analysis of this study. Two hundred volunteers under the age of 30years, who did not have a calcium score were excluded, as the calcium score was the main outcome of our analysis.

Subset analysis for activity tracker data was performed on 430 out of the 600 volunteers who had adequate data. Although subjects recruited were issued an activity tracker to be worn over a period of five days with first and last days of the study being partial days, there was inconsistent wearing of the activity. Discounting the partial days, each subject would yield an activity log for three complete tracking days (or equivalent to days with>20 valid hours of steps and sleep data)24,25. For data consistency and quality, subjects with improper activity tracker usage i.e. activity reading log less than five days and/or sleep reading log less than three days were censored.

Coronary artery calcium (CAC) scoring was used as the modelling outcome. The coronary calcium is a specific marker of coronary atherosclerosis, a precursor for coronary artery disease26; it also reflects arterial age under the influence of underlying comorbidities and lifestyle. The CAC score was also regarded as the best marker for risk prediction of cardiovascular events27,28.

This study stratified subjects into two classes of CVD risk. Low risk if their coronary artery calcium score were 0, and high risk if calcium score were 100 and above. Subjects who did not fall into these 2 categories were considered intermediate risk.

The aim of this study is to look at how accurate the machine learning algorithm is in handling different data types, in the task of predicting high risk and low risk patients, based on calcium score.

Table 1 summarizes the data from SingHEART that was used in this study.

Data variables were categorized into four groups; lifestyle survey questionnaires, blood test data, 24-h ambulatory blood pressure, and activity tracking data by commercially available Fitbit Charge HR29.

Data pre-processing, transformation and imputation were performed on the raw data. Variables selected were based on their a priori knowledge from previous publications on cardiovascular risk assessment1,2,3, and physician expert advice. In total, there were 30, 17, 12 and 16 unique variables in the respective groups: survey questionnaire, 24h blood pressure and heart rate monitoring, blood tests and Fitbit data.

The Framingham 10-year risk score was computed using seven traditional risk factors: gender, age, single timepoint systolic blood pressure, Total Cholesterol (TC), High Density Lipoprotein (HDL), smoking status and presence of diabetes. A Framingham risk score of<10% is consider low risk, while20% is considered high risk30.

Figure1 shows the methodological framework of the present study. Exploratory analysis showed that ensemble MLA classifiers were superior at discriminating low risk individuals while ensemble MLA regressors performed better identifying individuals with high CVD risk. To leverage on the merits of both the classifiers and regressors MLA, we used both approaches for our model.

Modelling flow chart using ensemble MLA for cardiovascular risk prediction.

The ensemble classifiers produce a binary prediction outcome; low or non-low risk. The ensemble regressors makes a numerical prediction on the calcium score for individuals classified as non-low risk, and stratify into three bins of low, high, and intermediate risk. The predicted numerical values may range from negative to positive number. Negative predicted values were first converted to zero and subsequently the continuous predictions were converted to discrete bins using unique value percentile discretization ensuring records with the same numerical prediction are assigned the same risk category. Finally, the prediction outcome resides in a decision node build on a rule-based logic. The decision node assigns an outcome of low risk if classifiers predict an individual to be low in CVD risk, high risk if classifier predicts non-low risk and regressor predicts high risk. Patients with incongruent classifiers and regressor outcomes are considered unclassified.

The ensemble models in both classification and regression phase each fit three base learners (naive bayes (NB), random forest (RF) and support vector classifier (SVC) for classification prediction, and generalized linear regression (GLM), support vector regressor (SVR) and stochastic gradient descent (SGD) for regression prediction). These base learners were chosen based on preliminary analysis, where these models showed efficiency in handling missing values and outliers.

The ensemble model then uses majority vote to determine the class label in classification phase. For the regression phase, the ensemble model averages the normalized predictions from the base regressor models to form a numerical outcome.

All models were trained on a stratified five-fold cross-validation. As SingHEART data had an imbalanced CVD risk distribution of risk based on the calcium score (low risk 63.4%, high risk 8.3%, intermediate risk 18.7%) we oversampled the training set for the minority class labels to allow model to better learn features from the under-represented classes31. The data were first partitioned into five mutually exclusive subsets, with each subset sharing the same proportion of class label as original dataset. At each iteration, the MLAs trained on four parts (80%) and validated on the fifth, the holdout set (20%). The process repeats five times, with five different but overlapping training sets. The resulting metrics from each fold were averaged to produce a single estimate.

To simulate access to the different variable groups as per clinical workflow and ease of information availability, we assessed the performance of individual variable group, and in combination as per the following:

Model 1: Survey Questionnaire.

Model 2: 24h ambulatory blood pressure and heart rate.

Model 3: Clinical blood results.

Model 4: Model 1+Model 2.

Model 5: Model 1+Model 3.

Model 6: Model 1 to Model 3.

Model 6*: Model 1 to Model 3 with feature selection.

Model 7: Physical activity and sleep trackers (exploratory subset analysis).

Variables in model 6* were reduced using SVC recursive feature elimination with cross-validation (SVC-RFECV) method to automatically select the best set of predictors that yield the highest area under Receiver Operating Characteristic curves (AUC). Model 16 were trained using 600 subjects.

We also performed exploratory analysis using MLA on the Fitbit Charge HR data (Model 7). Model 7 was trained on a subset of 430 subjects constrained by availability of valid activity tracking data.

Since no single metric can objectively evaluate the cardiovascular risk prediction, we evaluate the performance of our models at CVD risk class level using a panel of metrics; sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), F1-score and Area under Receiver Operating Characteristic curves (AUC). Overall discriminative ability of the model was described by the area under received operating characteristic curve (ROC). All AUC metrics were accompanied by 95% confidence interval (CI) and standard deviation (SD).

To better understand the relative importanceof different risk factors, we conduct a post-hoc approach to rank the variables by their contribution to CVD risk prediction. Feature importance were obtained from the SVC algorithm where the relative importance was determined by the absolute size of the coefficients in relation to others. All statistical analyses were conducted on Python version 3.7 environment and all MLAs and evaluation metrics were constructed using Scikit-learn libraries.

Continue reading here:
Application of ensemble machine learning algorithms on lifestyle factors and wearables for cardiovascular risk prediction | Scientific Reports -...

The 6 Best Deep Learning Tutorials on YouTube to Watch Right Now – Solutions Review

Learning deep learning can be a complicated process, and its not easy to know where to start. As a result, our editors have compiled this list of the best deep learning tutorials on YouTube to help you learn about the topic and hone your skills before you move on to mastering it. All of the videos here are free to access and feature guidance from some of the top minds and biggest brands in the online learning community. All of the best deep learning tutorials listed tout a minimum of 200,000 views.

Note: Dont forget to subscribe to Solutions Review on YouTube!

Author: Lex Fridman

Description: An introductory lecture for MIT course 6.S094 on the basics of deep learning including a few key ideas, subfields, and the big picture of why neural networks have inspired and energized an entirely new generation of researchers. The most popular result on YouTube touts nearly 1.2 million views.

Author: sentdex

Description: An updated deep learning introduction using Python, TensorFlow, and Keras, this tutorial tours more than a million views and is one of the most popular resources on the web. Students can access the text-tutorial and notes here, TensorFlow Docs here, and Keras docs here. There is also a community Discord server for those interested in the topic.

Author: Simplilearn

Description: This video provides a fun and simple introduction to deep learning concepts. Students learn about where deep learning is implemented and move on to how it is different from machine learning and artificial intelligence. Watchers will also look at what neural networks are and how they are trained to recognize digits written by hand.

Author: freeCodeCamp

Description: This course will teach you how to use Keras, a neural network API written in Python and integrated with TensorFlow. Students will learn how to prepare and process data for artificial neural networks, build and train artificial neural networks from scratch, build and train convolutional neural networks (CNNs), implement fine-tuning and transfer learning, and more.

Author: Edureka

Description: This Edureka deep learning full course video will help you understand and learn deep learning and TensorFlow are in detail. This deep learning tutorial is ideal for both beginners as well as professionals who want to master deep learning algorithms.

Author: Edureka

Description: This Edureka video will help you to understand the relationships between deep learning, machine learning, and artificial intelligence. This tutorial discusses AI, machine learning and its limitations, and how deep learning overcame machine learning limitations. Additional topics include deep learning applications and TensorFlow.

Author: freeCodeCamp

Description: Learn the fundamental concepts and terminology of seep learning, a sub-branch of machine learning. This course is designed for absolute beginners with no experience in programming. You will learn the key ideas behind deep learning without any code. It also covers neural networks and various machine learning constructs.

Tim is Solutions Review's Editorial Director and leads coverage on big data, business intelligence, and data analytics. A 2017 and 2018 Most Influential Business Journalist and 2021 "Who's Who" in data management and data integration, Tim is a recognized influencer and thought leader in enterprise business software. Reach him via tking at solutionsreview dot com.

Originally posted here:
The 6 Best Deep Learning Tutorials on YouTube to Watch Right Now - Solutions Review