Category Archives: Machine Learning
AI Weekly: The promise and limitations of machine programming tools – VentureBeat
Elevate your enterprise data technology and strategy at Transform 2021.
Machine programming, which automates the development and maintenance of software, is becoming supercharged by AI. During its Build developer conference in May, Microsoft detailed a new feature in Power Apps that taps OpenAIs GPT-3 language model to assist people in choosing formulas. Intels ControlFlag can autonomously detect errors in code. And Facebooks TransCoderconverts code from one programming language into another.
The applications of computer programming are vast in scope. And as computers become ubiquitous, the demand for quality code draws an ever-growing number of aspiring programmers to the profession. After years of study to become proficient at coding, experts learn to convert abstracts into concrete, executable programs. But they spend the majority of their work hours not programming. According to a study from the University of Cambridge, at least half of developers efforts are spent debugging, which costs the software industry an estimated $312 billion per year.
AI-powered code suggestion and review tools promise to cut development costs substantially while allowing coders to focus on more creative, less repetitive tasks, according to Justin Gottschlich, principal AI scientist at Intels machine programming division. Gottschlich is spearheading the work on ControlFlag, which fuses machine learning, formal methods, programming languages, and compilers to detect normal coding patterns, identifying abnormalities in code that are likely to cause a bug.
Prior to machine learning- or AI-based programming systems, programmers had dozens perhaps hundreds of tools to help them be more productive, produce code with fewer logic errors, improve the softwares performance, and so on. However, nearly all of these systems were rules-based,' Gottschlich told VentureBeat via email. While useful, rules-based systems are inherently limited in scope by the rules that they have been programmed into them. As such, if new kinds of things occur, the systems would need to be updated by humans. Moreover, these rules-based systems have always been prone to human error in creating the rules encoded in them. For example, programmers may accidentally create a rule to find a certain type of bug, but incorrectly define the rules to find it. This hidden bug in the rules system could go undetected forever.
Gottschlich asserts that AI-based systems offer benefits over the rules-based systems of yesteryear because AI can learn on its own in an unsupervised fashion, enabling it to draw on massive code databases. With unsupervised learning, an algorithm is fed unknown data for which no previously defined labels exist. The system must teach itself to classify the data by processing it to learn from its structure.
For example, ControlFlag was trained on over 1 billion unlabeled lines of code to identify stylistic variations in programming language. As for TransCoder, it learned to translate between C++, Java, and Python by analyzing a GitHub corpus containing over 2.8 million repositories. Microsoft trained a bug-spotting program on a dataset of 13 million work items and bugs from 47,000 developers across AzureDevOps and GitHub repositories. And code review platform DeepCodes algorithms were taught using billions of lines of code captured from public open source projects.
Theres a difference between AI-powered coding tools that can generate code from whole cloth versus augment a programmers workflow, of course. The latter is more common. Startups such as Tabine (formerly Codota) are developing platforms that suggest and autocomplete scripts in Python, C, HTML, Java, Scala, Kotlin, and JavaScript. Ponicode taps AI to check the accuracy of code. Intels Machine Inferred Code Similarity engine can determine when two pieces of code perform similar tasks, even when they use different structures and algorithms. And DeepCode offers a machine learning-powered system for whole-app code reviews as does Amazon.
Currently, we see a lot of AI-powered assistants, enabling software engineers to gain velocity and accuracy in their work. And the reason for the availability of more assistant tools than automation tools is that AI-powered automation has simply not yet reached the level of accuracy required, Ponicode CEO Patrick Joubert told VentureBeat. Our industry is still young, and even though we can already see the potential of automation with AI based code generators, we have to acknowledge that automatically generated code is still pretty unmaintainable and the overall quality is not meeting the right standards yet. While some engineers are working on the future of AI powered automation, my team and I, along with many other stakeholders, are dedicated to creating tools that can be used today. Within a few years I believe there will be enough tools to cover all steps of the development lifecycle.
For Joubert, the most intriguing categories of machine programming tools today are autocompletion and code analysis. Autocompletion systems like Tabnine and Kite employ AI to analyze semantics and make sense of code, autocompleting functions with a sense of the codes semantic content and purpose. As for code analysis tools like Snyk and DeepCode, theyre dedicated to finding vulnerabilities in the code and suggesting actions to resolve them often with surprising speed and precision.
When we see the numerous leaks and bugs from any software, including the ones built by leading multinationals, we can agree that [the software] industry has not yet matured. AI-powered coding tools are mostly meant to enhance the developer experience and empower them, thanks to greater velocity and greater efficiency, Joubert added. Behind these developer-focused benefits, I believe we are on the way to allowing software engineers to build industrial-grade software, where quality, innovation, and speed are reached systematically Autocompletion [in particular is] enabling software engineers to focus on the most complex part of their codebase and removing the burden of manually writing long strings of code.
Despite their potential, both AI-powered code generators and coding assistance tools have their limitations. For example, while GitHub has over 250 million code repositories alone, most of the data is unannotated. Theres only a few examples that describe precisely what the code does, posing a particular challenge for any system that cant learn from unlabeled data.
In an effort to address this, IBM recently released CodeNet, a 14-million-sample labeled dataset with 500 million lines of code written in 55 programming languages. The company claims that the rich annotations added to CodeNet make it suitable for a diverse set of tasks as opposed to other datasets specialized for specific programming tasks. Already, researchers at IBM have conducted several experiments with CodeNet, including code classification, code similarity evaluation, and code completion.
It is my speculation that in the next decade, code semantics understanding systems are likely to be one of the most important areas of machine programming in the coming decade, Joubert said. It depends on the domain the machine programming system is being applied to. For small programs, such as unit tests or regression tests, full program synthesizers are a reality today. Yet, for larger programs, its currently computationally intractable for machine programming systems to generate the potential thousands or millions of lines of code without the assistance of a programmer.
Boris Paskalev, the cofounder and CEO of DeepCode, calls creating a couple of lines of code with AI more of a toy than a productivity breakthrough. While techniques like natural language processing work well with text because theres fixed limits on the words and syntax that need to be understood, code isnt the same, he argues.
Since there are no formal rules for software development, [programming] is an art that requires a complete understanding of code and a developers intentions to produce something that works as expected without bugs, Paskalev told VentureBeat. As far as weve come in using machine learning and neural networks for code, were still only in the invention of the wheel phase machine learning is already proving to be very useful for code, but only after it goes through a semantic machine learning-representation of the code: making sure all semantic facts, variables, transitions, and logical interrelations are clearly represented and considered by the learning model.
To Paskalevs point, recent studies suggest that AI has a ways to go before it can reliably generate code. In June, a team of researchers at the University of California at Berkeley, Cornell, the University of Chicago, and the University of Illinois at Urbana-Champaign released APPS, a benchmark for code generation from natural language specifications. The team tested several types of models on APPS, including OpenAIs GPT-2, GPT-3, and an open source version of GPT-3 called GPT-Neo. In experiments, they discovered that the models could learn to generate code that solves easier problems but not without syntax errors. Approximately 59% of GPT-3s solutions for introductory problems had errors, while the best-performing model GPT-Neo attained only 10.15% accuracy.
When generating code from whole cloth, there are typically challenges around both specifying the intent and consuming the results, Tabine CEO Dror Weiss told VentureBeat. User intent can be specified in natural language by providing examples, writing code in a higher-level language, or in other means. But in most cases, this intent does not provide a full specification of the desired behavior. Also, the generated code may be following different route than what the developer had in mind. As such, it may be challenging for the developer to judge whether the code performs the desired operation exactly.
Facebook AI researchers Baptiste Rozire and Marie-Anne Lachaux, who worked on TransCoder, agree with Tabines assessment. It is inherently difficult to generate correct code from unspecific natural language problem descriptions that could correspond to several different code snippets. An easier task would be to generate code from an input that is more specific and closer to the output code, like pseudo-code or code written in a different language, they told VentureBeat. A huge obstacle to the adoption of methods generating large amounts of code without human supervision is that they would need to be extremely reliable to be used easily. Even a tool that could generate methods with 99% accuracy would fail to generate a working codebase of hundreds of functions. It could speedup the code generation process but would still require human testing and intervention.
Rozire and Lachaux also point out that tasks around code generation are generally much harder than classification tasks because the model has a lot of freedom and can create many different outputs, making it hard to control the correctness of the generation. Moreover, compared with natural languages, programming languages are very sensitive to small errors. A one-character difference can change the semantics of the code and make the output faulty.
Current machine learning algorithms may not be able to generalize well enough to different problems to match human performance for coding interviews without larger datasets or much better unsupervised pre-training methods, Rozire and Lachaux said.
Paskalev thinks itll be at least five to ten years until natural language processing enables developers to create meaningful components or even entire apps from a simple description. But Gottschlich is more optimistic. He notes that AI-powered coding tools arent just valuable in writing code, but also when it comes to lower-hanging fruit like upgrading existing code. Migrating an existing codebase to a modern or more efficient language like Java or C++, for example, requires expertise in both the source and target languages and its often costly. The Commonwealth Bank of Australia spent around $750 million over the course of five years to convert its platform from COBOL to Java.
Deep learning already enables us to cover the smaller tasks, the repetitive and redundant ones which clutter a software engineers routine. Today, AI can free software engineers from tedious tasks slowing them down and decreasing their creativity, Gottschlich said. The human mind remains far superior when it comes to creation, innovation, and designing the most complex parts of our softwares. Enabling them to increase velocity in these exciting, high added value parts of their work is, I believe, the most interesting way to leverage the power of machine learning today.
Joubert and Weiss say that the potential business value of machine programming also cant be ignored. An estimated 19% to 23% of software development projects fail, with that statistic holding steady for the past couple of decades. Standish Groupfound that challenged projects i.e., those that fail to meet scope, time, or budget expectations account for about 52% of software projects. Often, a lack of user involvement and clear requirements are to blame for missed benchmarks.
We see a great number of new tools using AI to enhance legacy code and help existing assets reach industrial-grade standards. We can elevate developer legacy code management workflows and be part of reducing the hefty level of technical debt built up over the past 50 years in the software industry, Joubert said. The days when developers had to write and read code line by line are gone. Im excited to see how the other steps in the software development lifecycle are going to be transformed and how tools will reach the same level that Kite or Snyk have attained. Leveraging AI to build efficient, one-purpose, tested, secure, and documented code effortlessly is going to profoundly change the way software companies can create incremental value and innovation.
From Weiss perspective, AI-powered coding tools can reduce costly interactions between developers like Q&A sessions and repetitive code review feedback while shortening the project onboarding process. [These] tools make all developers in the enterprise better. They take the collective code intelligence of the organization and make it available, during development time, to all developers. This allows any developer on the team to punch above their weight, he said.
For AI coverage, send news tips toKyle Wiggers and be sure to subscribe to the AI Weekly newsletterand bookmark our AI channel,The Machine.
Thanks for reading,
Kyle Wiggers
AI Staff Writer
See original here:
AI Weekly: The promise and limitations of machine programming tools - VentureBeat
Machine Learning Can Reduce Worry About Nanoparticles In Food – Texas A&M Today – Texas A&M University Today
Machine learning algorithms developed by researchers can predict the presence of any nanoparticle in most plant species.
Getty Images
While crop yield has achieved a substantial boost from nanotechnology in recent years, alarms over the health risks posed by nanoparticles within fresh produce and grains have also increased. In particular, nanoparticles entering the soil through irrigation, fertilizers and other sources have raised concerns about whether plants absorb these minute particles enough to cause toxicity.
In a new study published online in the journalEnvironmental Science and Technology,researchers at Texas A&M University have used machine learning to evaluate the salient properties of metallic nanoparticles that make them more susceptible for plant uptake. The researchers said their algorithm could indicate how much plants accumulate nanoparticles in their roots and shoots.
Nanoparticles are a burgeoning trend in several fields, including medicine, consumer products and agriculture. Depending on the type of nanoparticle, some have favorable surface properties, charge and magnetism, among other features. These qualities make them ideal for a number of applications. For example, in agriculture, nanoparticles may be used as antimicrobials to protect plants from pathogens. Alternatively, they can be used to bind to fertilizers or insecticides and then programmed for slow release to increase plant absorption.
These agricultural practices and others, like irrigation, can cause nanoparticles to accumulate in the soil. However, with the different types of nanoparticles that could exist in the ground and a staggeringly large number of terrestrial plant species, including food crops, it is not clearly known if certain properties of nanoparticles make them more likely to be absorbed by some plant species than others.
As you can imagine, if we have to test the presence of each nanoparticle for every plant species, it is a huge number of experiments, which is very time-consuming and expensive, said Xingmao Samuel Ma, associate professor in the Zachry Department of Civil and Environmental Engineering. To give you an idea, silver nanoparticles alone can have hundreds of different sizes, shapes and surface coatings, and so, experimentally testing each one, even for a single plant species, is impractical.
Instead, for their study, the researchers chose two different machine learning algorithms, an artificial neural network and gene-expression programming. They first trained these algorithms on a database created from past research on different metallic nanoparticles and the specific plants in which they accumulated. In particular, their database contained the size, shape and other characteristics of different nanoparticles, along with information on how much of these particles were absorbed from soil or nutrient-enriched water into the plant body.
Once trained, their machine learning algorithms could correctly predict the likelihood of a given metallic nanoparticle to accumulate in a plant species. Also, their algorithms revealed that when plants are in a nutrient-enriched or hydroponic solution, the chemical makeup of the metallic nanoparticle determines the propensity of accumulation in the roots and shoots. But if plants are grown in soil, the contents of organic matter and the clay in soil are key to nanoparticle uptake.
Ma said that while the machine learning algorithms could make predictions for most food crops and terrestrial plants, they might not yet be ready for aquatic plants. He also noted that the next step in his research would be to investigate if the machine learning algorithms could predict nanoparticle uptake from leaves rather than through the roots.
It is quiteunderstandable that people are concerned about the presence of nanoparticles in their fruits, vegetables and grains, said Ma. But instead of not using nanotechnology altogether, we would like farmers to reap the many benefits provided by this technology but avoid the potential food safety concerns.
Other contributors include Xiaoxuan Wang, Liwei Liu and Weilan Zhang from the civil and environmental engineering department.
This research is partly funded by the National Science Foundation and the Ministry of Science and Technology, Taiwan under the Graduate Students Study Abroad Program.
See the article here:
Machine Learning Can Reduce Worry About Nanoparticles In Food - Texas A&M Today - Texas A&M University Today
Akamai Unveils Machine Learning That Intelligently Automates Application and API Protections and Reduces Burden on Security Professionals – KPVI News…
CAMBRIDGE, Mass., June 16, 2021 /PRNewswire/ -- Akamai Technologies, Inc. (NASDAQ: AKAM), the world's most trusted solution for protecting and delivering digital experiences, today announces platform security enhancements to strengthen protection for web applications, APIs, and user accounts. Akamai's machine learning derives insight on malicious activity from more than 1.3 billion daily client interactions to intelligently automate threat detections, time-consuming tasks, and security logic to help professionals make faster, more trustworthy decisions regarding cyberthreats.
In its May 9 report Top Cybersecurity Threats in 2021, Forrester estimates that due to reasons "exacerbated by COVID-19 and the resulting growth in digital interactions, identity theft and account takeover increased by at least 10% to 15% from 2019 to 2020." The leading global research and advisory firm notes that we should "anticipate another 8% to 10% increase in identity theft and ATO [account takeover] fraud in 2021." With threat actors increasingly using automation to compromise systems and applications, security professionals must likewise automate defenses in parallel against these attacks to manage cyberthreats at pace.
New Akamai platform security enhancements include:
Adaptive Security Engine for Akamai's web application and API protection (WAAP) solutions, Kona Site Defender and Web Application Protector, is designed to automatically adapt protections with the scale and sophistication of attacks, while reducing the effort to maintain and tune policies. The Adaptive Security Engine combines proprietary anomaly risk scoring with adaptive threat profiling to identify highly targeted, evasive, and stealthy attacks. The dynamic security logic intelligently adjusts its defensive aggressiveness based on threat intelligence automatically correlated for each customer's unique traffic. Self-tuning leverages machine learning, statistical models, and heuristics to analyze all triggers across each policy to accurately differentiate between true and false positives.
Audience Hijacking Protection has been added to Akamai Page Integrity Manager to detect and block malicious activity in real time from client-side attacks using JavaScript, advertiser networks, browser plug-ins, and extensions that target web clients. Audience Hijacking Protection is designed to use machine learning to quickly identify vulnerable resources, detect suspicious behavior, and block unwanted ads, pop-ups, affiliate fraud, and other malicious activities aimed at hijacking your audience.
Bot Score and JavaScript Obfuscation have been added to Akamai Bot Manager, laying the foundation for ongoing innovations in adversarial bot management, including the ability to take action against bots aligned with corporate risk tolerance. Bot Score automatically learns unique traffic and bot patterns, and self-tunes for long-term effectiveness; JavaScript Obfuscation dynamically changes detections to prevent bot operators from reverse engineering detections.
Akamai Account Protector is a new solution designed to proactively identify and block human fraudulent activity like account takeover attacks. Using advanced machine learning, behavioral analytics, and reputation heuristics, Account Protector intelligently evaluates every login request across multiple risk and trust signals to determine if it is coming from a legitimate user or an impersonator. This capability complements Akamai's bot mitigation to provide effective protection against both malicious human actors and automated threats.
"At Akamai, our latest platform release is intended to help resolve the tension between security and ease of use, with key capabilities around automation and machine learning specifically designed to intelligently augment human decision-making," said Aparna Rayasam, senior vice president and general manager, Application Security, Akamai. "Smart automation adds immediate value and empowers users with the right tools to generate insight and context to make faster and more trustworthy decisions, seamlessly all while anticipating what attackers might do next."
For more information about Akamai's Edge Security solutions, visit our Platform Update page.
About Akamai
Akamai secures and delivers digital experiences for the world's largest companies. Akamai's intelligent edge platform surrounds everything, from the enterprise to the cloud, so customers and their businesses can be fast, smart, and secure. Top brands globally rely on Akamai to help them realize competitive advantage through agile solutions that extend the power of their multi-cloud architectures. Akamai keeps decisions, apps, and experiences closer to users than anyone and attacks and threats far away. Akamai's portfolio of edge security, web and mobile performance, enterprise access, and video delivery solutions is supported by unmatched customer service, analytics, and 24/7/365 monitoring. To learn why the world's top brands trust Akamai, visit http://www.akamai.com, blogs.akamai.com, or @Akamai on Twitter. You can find our global contact information at http://www.akamai.com/locations.
Contacts:
Tim Whitman
Media Relations
617-444-3019
Tom Barth
Investor Relations
617-274-7130
View original content to download multimedia:http://www.prnewswire.com/news-releases/akamai-unveils-machine-learning-that-intelligently-automates-application-and-api-protections-and-reduces-burden-on-security-professionals-301313433.html
SOURCE Akamai Technologies, Inc.
Read more from the original source:
Akamai Unveils Machine Learning That Intelligently Automates Application and API Protections and Reduces Burden on Security Professionals - KPVI News...
Using large-scale experiments and machine learning to discover theories of human decision-making – Science Magazine
Discovering better theories
Theories of human decision-making have proliferated in recent years. However, these theories are often difficult to distinguish from each other and offer limited improvement in accounting for patterns in decision-making over earlier theories. Peterson et al. leverage machine learning to evaluate classical decision theories, increase their predictive power, and generate new theories of decision-making (see the Perspective by Bhatia and He). This method has implications for theory generation in other domains.
Science, abe2629, this issue p. 1209; see also abi7668, p. 1150
Predicting and understanding how people make decisions has been a long-standing goal in many fields, with quantitative models of human decision-making informing research in both the social sciences and engineering. We show how progress toward this goal can be accelerated by using large datasets to power machine-learning algorithms that are constrained to produce interpretable psychological theories. Conducting the largest experiment on risky choice to date and analyzing the results using gradient-based optimization of differentiable decision theories implemented through artificial neural networks, we were able to recapitulate historical discoveries, establish that there is room to improve on existing theories, and discover a new, more accurate model of human decision-making in a form that preserves the insights from centuries of research.
See more here:
Using large-scale experiments and machine learning to discover theories of human decision-making - Science Magazine
Data Insights and Machine Learning Take Charge of the Maritime Sales Process – Hellenic Shipping News Worldwide
While the maritime industry has been hesitant engaging in use of data insight and machine learning, the table is now about to turn. Today, an increasing number of maritime companies actively use data insights to improve sales, supply chain activities, and increase revenues among these the worlds largest ship supplier, Wrist Ship Supply.
The need for efficiency in the maritime sector has led companies to actively use data as a measure to optimize the supply chain. This has paved the way for new ship supply services centered around data insights and machine learning to increase top and bottom-line figures.
According to the leading data and analytics firm, GateHouse Maritime, data insights can make a noticeable difference in the maritime sector. With a combination of historic and real-time ocean data, machine learning, and smart algorithms, maritime supply companies can predict vessel destinations and arrivals with high precision.
Traditionally, vessel tracking has been a time consuming, manual process characterized by imprecise predictions and uncertainty. But today, the process can be automated and turn large amounts of data into tangible leads and sales:
With the help of data insights, it is possible to predict arrivals several days in advance with almost 100 percent accuracy. This allows maritime supply companies to obtain an obvious competitive advantage, as they can operate proactively and sell services to potential customers days before a given vessel calls into port, says CEO at GateHouse, Maritime, Martin Dommerby Kristiansen.
Data analytics strengthen the worlds largest ship supplierFour years ago, the worlds largest ship supplier, Wrist Ship Supply, realized a strategy that would integrate data analytics in numerous business areas. The global ship supplier is a full-service provider, providing service for marine, offshore and navy operations, such as supplying consumables, handling of owners goods and spare parts storage and forwarding.
Today, Wrist Ship Supply works strategically with data analytics and business intelligence to improve internal processes and increase value for customers:
In recent years, we have experienced an increasing pull from the market and as a market leader within ship supply, we feel obliged to take part in the digital transformation. Data analysis has proven to be a cornerstone and a very important tool for measuring and improving performances across our own as well as customers supply chain. Now, our business model is infused with data analytics and business intelligence that strengthen efficiency and reliability in both internal and external operations, explains Business Analysis Director at Wrist Ship Supply, Birthe Boysen.
For Birthe Boysen and Wrist Ship Supply, data analytics has especially proven its worth within sales:
It is crucial for us to know where potential customer vessels are heading and when they arrive in different ports. This allows us to coordinate our sales efforts and establish contact in advance. Not only does this make us more efficient, but it also creates value for customers, because all service activities can be planned several days ahead of arrival.
While the data-driven sales approach has increased the focus on KPIs, it has also become an important part of budgeting. Therefore, it has been a key priority for Wrist Ship Supply to be able to navigate in the ocean of available data:
We have an almost endless amount of data available, and it easily becomes exhausting to keep track of numbers and figures. Therefore, we prioritize to make sure that both internal and external stakeholders can make sense of the conclusions in our data insights. If employees or customers cannot fathom the overall lines in our data results, it will be difficult to use analytics in any way, Nadia Hay Kragholm, Senior Business Analyst in Wrist remarks.
According to Martin Dommerby Kristiansen, data insight has the potential to transform the entire maritime industry because efficiency has never been more important:
The maritime industry is indeed reliant on efficiency across the value chain. Recently, we have seen how a vessel stuck in the Suez Canal for only a few days can impact not only the maritime industry, but the entire transportation and logistics sector. This goes to show how important data insight and analytics can prove to be for companies that wish to operate proactively and minimize disorder in the supply chain.
GateHouse Maritime is a leader in Ocean Visibility solutions. We help global maritime service providers, cargo owners and logistic companies with transparent and accurate location data and predictions, cargo transport status, and offshore asset protection and surveillance. Our powerful maritime data foundation consists of 273 billion datapoints and +30 analysis and predictive models used for data-driven decisions by maritime operators worldwide. GateHouse Maritime is a subsidiary of GateHouse Holding, founded in 1992 and headquartered in Denmark, and which also holds the subsidiaries GateHouse SatCom and GateHouse Igniter.Source: GateHouse Maritime A/S
Go here to read the rest:
Data Insights and Machine Learning Take Charge of the Maritime Sales Process - Hellenic Shipping News Worldwide
Mydecine Innovations kicks off machine learning-based drug discovery program with the University of Alberta – Proactive Investors USA & Canada
The program will enable the company to more rapidly screen hundreds of thousands of new molecules without the need to produce them, allowing Mydecine to focus on the strongest potential therapeutics
() () () has launched its in-silico drug discovery program in conjunction with researchers at the University of Alberta (UofA), the company announced.
Led by computer-assisted drug development expert and UofA assistant professor at the Li Ka Shing Institute of Virology, Khaled Barakat, the program is focused on utilizing artificial intelligence/machine learning (AI/ML) to support drug screenings, including both the ability to build drugs from the receptor up and assess drugs around the receptors of Mydecines choosing.
The in-silico (read: computer simulated) program will enable the company to more rapidly screen hundreds of thousands of new molecules without the need to produce them, allowing Mydecine to focus on the strongest potential therapeutics for its chemical and natural development programs, the company said.
Mydecine will also be able to more efficiently screen its own proprietary library of novel compounds designed by Chief Science Officer Rob Roscow and advisory board member, Denton Hoyer.
Years of research have shown that the chemical components of psychoactive and non-psychoactive mushrooms can be extremely powerful in a therapeutic setting and yet, there is still so much that we dont understand about how these molecules can affect biological systems, CEO Josh Bartch said in a statement.
As the next evolution of drug discovery progresses forward, we strongly believe that this new age will be fully led by artificial intelligence and machine learning. Expanding our R&D efforts with the addition of our cutting-edge AI/ML drug screening program will allow our research teams to take a leading role within the psychedelic community to more efficiently expand our knowledge of these components and their pharmacological value.
At UofA, Barakat and his team specialize in understanding the nature and biophysical processes underlying protein-drug interaction, protein-protein interactions, protein-DNA interactions, drug off-target interactions and predicting drug-mediated toxicity.
Dr. Barakat and his team have built an impressive reputation as leaders at the intersection of technology and pharmacological science, Bartch said. Adding their specialization in developing innovative computer models and novel technologies to predict protein-protein and protein-drug interactions will bring tremendous value to Mydecines research and enable us to more quickly bring to market effective drugs that can produce better outcomes for patients.
Contact Andrew Kessel at andrew.kessel@proactiveinvestors.com
Follow him on Twitter @andrew_kessel
Originally posted here:
Mydecine Innovations kicks off machine learning-based drug discovery program with the University of Alberta - Proactive Investors USA & Canada
2 supervised learning techniques that aid value predictions – TechTarget
This article is excerpted from the course "Fundamental Machine Learning," part of the Machine Learning Specialist certification program from Arcitura Education. It is the ninth part of the 13-part series, "Using machine learning algorithms, practices and patterns."
This article explores the numerical prediction and category prediction supervised learning techniques. These machine learning techniques are applied when the target whose value needs to be predicted is known in advance and some sample data is available to train a model. As explained in Part 4, these techniques are documented in a standard pattern profile format.
A data set may contain a number of historical observations (rows) amassed over a period of time where the target value is numerical in nature and is known for those observations. An example is the number of ice creams sold and the temperature readings, where the number of ice creams sold is the target variable. To obtain value from this data, a business use case might require a prediction of how much ice cream will be sold if the temperature reading is known in advance from the weather forecast. As the target is numerical in nature, supervised learning techniques that work with categorical targets cannot be applied (Figure 1).
The historical data is capitalized upon by first finding independent variables that influence the target dependent variable and then quantifying this influence in a mathematical equation. Once the mathematical equation is complete, the value of the target variable is predicted by inputting the values of the independent values.
The data set is first scanned to find the best independent variables by applying the associativity computation pattern to find the relationship between the independent variables and the dependent variable. Only the independent variables that are highly correlated with the dependent variable are kept. Next, linear regression is applied.
Linear regression, also known as least squares regression, is a statistical technique for predicting the values of a continuous dependent variable based on the values of an independent variable. The dependent and independent variables are also known as response and explanatory variables, respectively. As a mathematical relationship between the response variable and the explanatory variables, linear regression assumes that a linear correlation exists between the response and explanatory variables. A linear correlation between response and explanatory variables is represented through the line of best fit, also called a regression line. This is a straight line that passes as closely as possible through all points on the scatter plot (Figure 2).
Linear regression model development starts by expressing the linear relationship. Once the mathematical form has been established, the next step is to estimate the parameters of the model via model fitting. This determines the line of best fit achieved via least squares estimation that aims to reduce the sum of squared error (SSE). The last stage is to evaluate the model either using R squared or mean squared error (MSE).
MSE is a measure that determines how close the line of best fit is to the actual values of the response variable. Being a straight line, the regression line cannot pass through each point; it is an approximation of the actual value of the response variable based on estimated values. The distance between the actual and the estimated value of response variable is the error of estimation. For the best possible estimate of the response variable, the errors between all points, as represented by the sum of squared error, must be minimized. The line of best fit is the line that results in the minimum possible sum of squares errors. In other words, MSE identifies the variation between the actual value and the estimated value of the response variable as provided by the regression line (Figure 3).
The coefficient of determination, called R squared, is the percentage of variation in the response variable that is predicted or explained by the explanatory variable, with values that vary between 0 and 1. A value equal to 0 means that the response variable cannot be predicted from the explanatory variable, while a value equal to 1 means the response variable can be predicted without any errors. A value between 0 and 1 provides the percentage of successful prediction.
In regression, more than two explanatory variables can be used simultaneously for predicting the response variable, in which case it is called multiple linear regression.
The numerical prediction pattern can benefit from the application of the graphical summaries computation pattern by drawing a scatter plot to graphically validate if a linear relationship exists between the response and explanatory variables (Figure 4).
There are cases where a business problem involves predicting a category -- such as whether a customer will default on their loan or whether an image is a cat or a dog -- based on historical examples of defaulters and cats and dogs, respectively. In this case, the categories (default/not default and cat/dog) are known in advance. However, as the target class is categorical in nature, numerical predictive algorithms cannot be applied to train and predict a model for classification purposes (Figure 5).
Supervised machine learning techniques are applied by selecting a problem-specific machine learning algorithm and developing a classification model. This involves first using the known example data to train a model. The model is then fed new unseen data to find out the most appropriate category to which the new data instance belongs.
Different machine learning algorithms exist for developing classification models. For example, naive Bayes is probabilistic while K-nearest neighbors (KNN), support vector machine (SVM), logistic regression and decision trees are deterministic in nature. Generally, in the case of a binary problem -- cat or dog -- logistic regression is applied. If the feature space is n-dimensional (a large number of features) with complex interactions between the features, KNN is applied. Naive Bayes is applied when there is not enough training data or fast predictions are required, while decision trees are a good choice when the model needs to be explainable.
Logistic regression is based on linear regression and is also considered a class probability estimation technique, since its objective is to estimate the probability of an instance belonging to a particular class.
KNN, also known as lazy learning and instance-based learning, is a black-box classification technique where instances are classified based on their similarity, with a user-defined (K) number of examples (nearest neighbors). No model is explicitly generated. Instead, the examples are stored as-is and an instance is classified by first finding the closest K examples in terms of distance, then assigning the class based on the class of the majority of the closest examples (Figure 6).
Naive Bayes is a probability-based classification technique that predicts class membership based on the previously observed probability of all potential features. This technique is used when a combination of a number of features, called evidence, affects the determination of the target class. Due to this characteristic, naive Bayes can take into account features that may be insignificant when considered on their own but when considered accumulatively can significantly impact the probability of an instance belonging to a certain class.
All features are assumed to carry equal significance, and the value of one feature is not dependent on the value of any other feature. In other words, the features are independent. It serves as a baseline classifier for comparing more complex algorithms and can also be used for incremental learning, where the model is updated based on new example data without the need for regenerating the whole model from scratch.
A decision tree is a classification algorithm that represents a concept in the form of a hierarchical set of logical decisions with a tree-like structure that is used to determine the target value of an instance. [See discussion of decision trees in part 2 of this series.] Logical decisions are made by performing tests on the feature values of the instances in such a way that each test further filters the instance until its target value or class membership is known. A decision tree resembles a flowchart consisting of decision nodes, which perform a test on the feature value of an instance, and leaf nodes, also known as terminal nodes, where the target value of the instance is determined as a result of traversal through the decision nodes.
The category prediction pattern normally requires the application of a few other patterns. In the case of logistic regression and KNN, applying the feature encoding pattern ensures that all features are numerical as these two algorithms only work with numerical features. The application of the feature standardization pattern in the case of KNN ensures that none of the large magnitude features overshadow smaller magnitude features in the context of distance measurement. Naive Bayes requires the application of the feature discretization pattern as naive Bayes only works with nominal features. KNN can also benefit from the application of feature discretization pattern via a reduction in feature dimensionality, which contributes to faster execution and increased generalizability of the model.
The next article covers the category discovery and pattern discovery unsupervised learning patterns.
Read the original post:
2 supervised learning techniques that aid value predictions - TechTarget
The CIO’s Guide to Building a Rockstar Data Science and AI Team | eWEEK – eWeek
Just about everyone agrees that data scientists and AI developers are the new superstars of the tech industry. But ask a group of CIOs to define the precise area of expertise for data science-related job titles, and discord becomes the word of the day.
As businesses seek actionable insights by hiring teams that include data analysts, data engineers, data scientists, machine learning engineers and deep learning engineers, a key to success is understanding what each role can and cant do for the business.
Read on to learn what your data science and AI experts can be expected to contribute as companies grapple with ever-increasing amounts of data that must be mined to create new paths to innovation.
In a perfect world, every company employee and executive works under a well-defined set of duties and responsibilities.
Data science isnt that world. Companies often will structure their data science organization based on project need: Is the main problem maintaining good data hygiene? Or is there a need to work with data in a relational model? Perhaps the team requires someone to be an expert in deep learning, and to understand infrastructure as well as data?
Depending on a companys size and budget, any one job title might be expected to own one or more of these problem-solving skills. Of course, roles and responsibilities will change with time, just as theyve done as the era of big data evolves into the age of AI.
That said, its good for a CIO and the data science team she or he is managing today to remove as much of the ambiguity as possible regarding roles and responsibilities for some of the most common roles those of the data analyst, data engineer, data scientist, machine learning engineer and deep learning engineer.
Teams that have the best understanding of how each fits into the companys goals are best positioned to deliver a successful outcome. No matter the role, accelerated computing infrastructure is also key to powering success throughout the pipeline as data moves from analytics to advanced AI.
Its important to recognize the work of a data analyst, as these experts have been helping companies extract information from their data long before the emergence of the modern data science and AI pipeline.
Data analysts use standard business intelligence tools like Microsoft Power BI, Tableau, Qlik, Yellowfin, Spark, SQL and other data analytics applications. Broad-scale data analytics can involve the integration of many different data sources, which increases the complexity of the work of both data engineers and data scientists another example of how the work of these various specialists tends to overlap and complement each other.
Data analysts still play an important role in the business, as their work helps the business assess its success. A data engineer might also support a data analyst who needs to evaluate data from different sources.
Data scientists take things a step further so that companies can start to capitalize on new opportunities with recommender systems, conversational AI, and computer vision, to name a few examples.
A data engineer makes sense of messy data and theres usually a lot of it. People in this role tend to be junior teammates who make data nice and neat (as possible) for data scientists to use. This role involves a lot of data prep and data hygiene work, including lots of ETL (extract, transform, load) to ingest and clean data.
The data engineer must be good with data jigsaw puzzles. Formats change, standards change, even the fields a team is using on a webpage can change frequently. Datasets can have transmission errors, such as when data from one field is incorrectly entered into another.
When datasets need to be joined together, data engineers need to fix the data hygiene problems that occur when labeling is inconsistent. For example, if the day of the week is included in the source data, the data engineer needs to make sure that the same format is used to indicate the day, as Monday could also be written as Mon., or even represented by a number that could be one or zero depending on how the days of the week are counted.
Expect your data engineers to be able to work freely with scripting languages like Python, and in SQL and Spark. Theyll need programming language skills to find problems and clean them up. Given that theyll be working with raw data, their work is important to ensuring your pipeline is robust.
If enterprises are pulling data from their data lake for AI training, this rule-based work can be done by a data engineer. More extensive feature engineering is the work of a data scientist. Depending on their experience and the project, some data engineers may support data scientists with initial data visualization graphs and charts.
Depending on how strict your company has been with data management, or if you work with data from a variety of partners, you might need a number of data engineers on the team. At many companies, the work of a data engineer often ends up being done by a data scientist, who preps her or his own data before putting it to work.
Data scientists experiment with data to find the secrets hidden inside. Its a broad field of expertise that can include the work of data analytics and data processing, but the core work of a data scientist is done by applying predictive techniques to data using statistical machine learning or deep learning.
For years, the IT industry has talked about big data and data lakes. Data scientists are people who finally turn these oceans of raw data into information. These experts use a broad range of tools to conduct analytics, experiment, build and test models to find patterns. To be great at their work, data scientists also need to understand the needs of the business theyre supporting.
These experts use many applications, including NumPy, SciKit-Learn, RAPIDS, CUDA, SciPy, Matplotlib, Pandas, Plotly, NetworkX, XGBoost, domain-specific libraries and many more. They need to have domain expertise in statistical machine learning, random forests, gradient boosting, packages, feature engineering, training, model evaluation and refinement, data normalization and cross-validation. The depth and breadth of these skills make it readily apparent why these experts are so highly valued at todays data-driven companies.
Data scientists often solve mysteries to get to the deeper truth. Their work involves finding the simplest explanations for complex phenomena and building models that are simple enough to be flexible yet faithful enough to provide useful insight. They must also avoid some perils of model training, including overfitting their data sets (that is, producing models that do not effectively generalize from example data) and accidentally encoding hidden biases into their models.
A machine learning engineer is the jack of all trades. This expert architects the entire process of machine and deep learning. They take AI models developed by data scientists and deep learning engineers and move them into production.
These unicorns are among the most sought-after and highly paid in the industry and companies work hard to make sure they dont get poached. One way to keep them happy is to provide the right accelerated computing resources to help fuel their best work. A machine learning engineer has to understand the end-to-end pipeline, and they want to ensure that pipeline is optimized to deliver great results, fast.
Its not always easily intuitive, as the machine learning engineers must know the apps, understand the downstream data architecture, and key in on system issues that may arise as projects scale. A person in this role must understand all the applications used in the AI pipeline, and usually needs to be skilled in infrastructure optimization, cloud computing, containers, databases and more.
To stay current, AI models need to be reevaluated to avoid whats called model drift as new data impacts the accuracy of the predictions. For this reason, machine learning engineers need to work closely with their data science and deep learning colleagues who will need to reassess models to maintain their accuracy.
A critical specialization for the machine learning engineer is deep learning engineer. This person is a data scientist who is an expert in deep learning techniques. In deep learning, AI models are able to learn and improve their own results through neural networks that imitate how human beings think and learn.
These computer scientists specialize in advanced AI workloads. Their work is part science and part art to develop what happens in the black box of deep learning models. They do less feature engineering and far more math and experimentation. The push for explainable AI (XAI) model interpretability and explainability can be especially challenging in this domain.
Deep learning engineers will need to process large datasets to train their models before they can be used for inference, where they apply what theyve learned to evaluate new information. They use libraries like PyTorch, TensorFlow and MXNet, and need to be able to build neural networks and have strong skills in statistics, calculus and linear algebra.
Given all of the broad expertise in these key roles, its clear that enterprises need a strategy to help them grow their teams success in data science and AI. Many new applications need to be supported, with the right resources in place to help this work get done as quickly as possible to solve business challenges.
Those new to data science and AI often choose to get started with accelerated computing in the cloud, and then move to a hybrid solution to balance the need for speed with operational costs. In-house teams tend to look like an inverted pyramid, with more analysts and data engineers funneling data into actionable tasks for data scientists, up to the machine learning and deep learning engineers.
Your IT paradigm will depend on your industry and its governance, but a great rule of thumb is to ensure your vendors and the skills of your team are well aligned. With a better understanding of the roles of a modern data team, and the resources they need to be successful, youll be well on your way to building an organization that can transform data into business value.
ABOUT THE AUTHOR
By Scott McClellan, Head of Data Science, NVIDIA
Read more here:
The CIO's Guide to Building a Rockstar Data Science and AI Team | eWEEK - eWeek
AWS leader talks about technologies needed to take precision medicine to the next level – Healthcare IT News
One of the most significant challenges to the advancement of precision medicine has been the lack of an infrastructure to support translational bioinformatics, supporting organizations as they work to uncover unique datasets to find novel associations and signals.
By supporting greater interoperability and collaboration, data scientists, developers, clinicians and pharmaceutical partners have the opportunity to leverage machine learning to reduce the time it takes to move from insight to discovery, ultimately leading to the right patients receiving the right care, with the right therapeutic at the right time.
To get a better understanding of challenges surrounding precision medicine and its future, Healthcare IT News sat down with Dr. Taha Kass-Hout, director of machine learning at AWS.
Q: You've said that one of the most significant challenges to the advancement of precision medicine has been the lack of an infrastructure to support translational bioinformatics. Please explain this challenge in detail.
A: One of the challenges in developing and utilizing storage, analytics and interpretive methods is the sheer volume of biomedical data that needs to be transformed that often resides on multiple systems and in multiple formats. The future of healthcare is so vibrant and dynamic and there is an opportunity for cloud and big data to take on a larger role to help the industry address these areas.
For example, datasets used to perform tasks such as computational chemistry and molecular simulations that help de-risk, and advance molecules into development, contain millions of data points and require billions of calculations to produce an experimental output. In order to bring new therapeutics to market faster, scientists need to move targets through development faster and find more efficient ways to collaborate both inside and outside of their organizations.
Another challenge is that large volumes of data acquired by legacy research equipment, such as microscopes and spectrometers, is usually stored locally. This creates a barrier for securely archiving, processing and sharing with collaborating researchers globally. Improving access to data, securely and compliantly, while increasing usability is critical to maximizing the opportunities to leverage analytics and machine learning.
For instance, Dotmatics' cloud-based software provides simple, unified, real-time access to all research data in Dotmatics and third-party databases, coupled with integrated, scientifically aware informatics solutions for small molecule and biologics discovery that expedite laboratory workflows and capture experiments, entities, samples and test data so that in-house or multi-organizational research teams become more efficient.
Today we are seeing a rising wave of healthcare organizations moving to the cloud, which is enabling researchers to unite R&D data with information from across the value chain, while benefiting from compute and storage options that are more cost-effective than on-premises infrastructure.
For large datasets in the R&D phase, large-scale, cloud-based data transfer services can transfer hundreds of terabytes and millions of files at speeds up to 10 times faster than open-source tools. Storage gateways ensure experimental data is securely stored, archived and available to other permissioned collaborators. Uniting data in a data lake improves access and helps to eliminate silos.
Cloud-based hyperscale computing and machine learning enable organizations to collaborate across datasets, create and leverage global infrastructures to maintain data integrity, and more easily perform machine learning-based analyses to accelerate discoveries and de-risk candidates faster.
For example, six years agoModerna started building databases and information-based activities to support all of their programs. Today, they are fully cloud-based, and their scientists don't go to the lab to pipette their messenger RNA and proteins. They go to their web portal, the Drug Design Studio that runs on the cloud.
Through the portal, scientists can access public and private libraries that contain all the messenger RNA that exists and the thousands of proteins they can produce. Then, they only need to press a button and the sequence goes to a fully automated, central lab where data is collected at every step.
Over the years, data from the portal and lab has helped Moderna improve their sequence design and production processes and improve the way their scientists gather feedback. In terms of research, all of Moderna's algorithms rely on computational power from the cloud to further their science.
Q: You contend that by supporting greater interoperability and collaboration, data scientists, developers, clinicians and pharmaceutical partners have the opportunity to leverage machine learning to reduce the time it takes to move from insight to discovery. Please elaborate on machine learning's role here in precision medicine.
A: For the last decade, organizations have focused on digitizing healthcare. In the next decade, making sense of all this data will provide the biggest opportunity to transform care. However, this transformation will primarily depend on data flowing where it needs to, at the right time, and supporting this process in a way that is secure and protects patients' health data.
It comes down to interoperability. It may not be the most exciting topic, but it's by far one of the most important, and one the industry needs to prioritize. By focusing on interoperability of information and systems today, we can ensure that we end up in a better place in 10 years than where we are now. And so, everything around interoperability around security, around identity management, differential privacy is likely to be part of this future.
Machine learning models trained to support healthcare and life sciences organizations can help automatically normalize, index and structure data. This approach has the potential to bring data together in a way that creates a more complete view of a patient's medical history, making it easier for providers to understand relationships in the data and compare this to the rest of the population, drive increased operational efficiency, and have the ability to use data to support better patient health outcomes.
For example, AstraZeneca has been experimenting with machine learning across all stages of research and development, and most recently in pathology to speed up the review of tissue samples. Labeling the data is a time-consuming step, especially in this case, where it can take many thousands of tissue-sample images to train an accurate model.
AstraZeneca uses a machine learning-powered, human-in-the-loop data-labeling and annotation service to automate some of the most tedious portions of this work, resulting in at least 50% less time spent cataloging samples.
It also helps analysts spot trends and anomalies in the health data and derive actionable insights to improve the quality of patient care, make predictions for medical events such as stroke or congestive heart failure, modernize care infrastructure, increase operational efficiency and scale specialist expertise.
Numerate, a discovery-stage pharmaceutical, uses machine learning technologies to more quickly and cost-effectively identify novel molecules that are most likely to progress through the research pipeline and become good candidates for new drug development.
The company recently used its cloud-based platform to rapidly discover and optimize ryanodine receptor 2 (RYR2) modulators, which are being advanced as new drugs to treat life-threatening cardiovascular diseases.
Ryanodine 2 is a difficult protein to target, but the cloud made that process easier for the company. Traditional methods could not have attacked the problem, as the complexity of the biology makes the testing laborious and slow, independent of the industry's low 0.1% screening hit rate for much simpler biology.
In Numerate's case, using the cloud enabled the company to effectively decouple the trial-and-error process from the laboratory and discover and optimize candidate drugs five times faster than the industry average.
Machine learning also is helping power the entire clinical development process. Biopharma researchers use machine learning to design the most productive trial protocols, study locations, recruitmentand patient cohorts to enroll. Researchers not trained as programmers can use cloud-based machine learning services to build, train and deploy machine learning algorithms to help with pre-clinical studies, complex simulations and predictive workflow optimization.
Machine learning can also help accelerate the regulatory submission process, as the massive amounts of data generated during clinical trials can be captured and effectively shared to collaborate between investigators, contract research organizations (CROs) and sponsor organizations.
For example, the Intelligent Trial Planner (ITP) from Knowledgent, now part of Accenture, uses machine learning services to determine the feasibility of trial studies and forecast recruitment timelines. The ITP platform enables study design teams at pharma organizations to run prediction analysis in minutes, not weeks, allowing them to iterate faster and more frequently.
Powered by machine learning, real-time scenario planning helps to facilitate smarter trial planning by enabling researchers to determine the most optimal sites, countries and/or protocol combinations.
By eliminating poor performing sites, trial teams have the potential to reduce their trial cost by 20%. And by making data-driven decisions that are significantly more accurate, they can plan and execute clinical trials faster, leading to hundreds of thousands in cost savings for every month saved in a trial.
Additionally, purpose-built machine learning is supported by cost-effective cloud-based compute options. For example, high-performance computing (HPC) can quickly scale to accommodate large R&D datasets, orchestrating services and simplifying the use and management of HPC environments.
Data transformation tools can also help to simplify and accelerate data profiling, preparation and feature engineering, as well as enable reusable algorithms both for new model discovery and inference.
The healthcare and life sciences industry has come a long way in the last year. However, for progress and transformation to continue, interoperability needs to be prioritized.
Q: The ultimate goal of precision medicine is the right patients receiving the right care, with the right therapeutic, at the right time. What do healthcare provider organization CIOs and other health IT leaders need to be doing with machine learning and other technologies today to be moving toward this goal?
A: The first things IT leaders need to ask themselves is: 1) If they are not investing yet in machine learning, do they plan to this year? And 2) What are the largest blockers to machine learning in their teams?
Our philosophy is to make machine learning available to every data scientist and developer without the need to have a specific background in machine learning, and then have the ability to use machine learning at scale and with cost efficiencies.
Designing a personalized care pathway using therapeutics tuned for particular biomarkers relies on a combination of different data sources such as health records and genomics to deliver a more complete assessment of a patient's condition. By sequencing the genomes of entire populations, researchers can unlock answers to genetic diseases that historically haven't been possible in smaller studies and pave the way for a baseline understanding of wellness.
Population genomics can improve the prevention, diagnosis and treatment of a range of illnesses, including cancer and genetic diseases, and produce the information doctors and researchers need to arrive at a more complete picture of how an individual's genes influence their health.
Advanced analytics and machine learning capabilities can use an individual or entire population's medical history to better understand relationships in data and in turn deliver more personalized and curated treatment.
Second, healthcare and life sciences organizations need to be open to experimenting, learning about and embracing both cloud and technology and many organizations across the industry are already doing this.
Leaders in precision medicine research such as UK Biobank, DNAnexus, Genomics England, Lifebit, Munich Lukemia Lab, Illumina, Fabric Genomics, CoFactor Genomics and Emedgene all leverage cloud and technology to speed genomic interpretation.
Third, supporting open collaboration and data sharing needs to be a business priority. The COVID-19 Open Research Dataset (CORD-19) created last year by a coalition of research groups provided open access to the plenary of available global COVID-19 research and data.
This was one of the primary factors that enabled the discovery, clinical trial and delivery of the mRNA-based COVID-19 vaccines in an unprecedented timeframe. Additionally, our Open Data Programmakes more than 40 openly available genomics datasets accessible, providing the research community with a single documented source of truth.
Commercial solutions that have leveraged machine learning to enable large-scale genomic sequencing include organizations such as Munich Leukemia Lab, who has been able to use the Field Programmable Gate Array-based compute instances to greatly speed up the process of whole genome sequencing.
As a result, what used to take 20 hours of compute time can now be achieved in only three hours. Another example is Illumina, which is using cloud solutions to offer its customers a lower-cost, high-performance genomic analysis platform, which can help them speed their time to insights as well as discoveries.
Twitter:@SiwickiHealthITEmail the writer:bsiwicki@himss.orgHealthcare IT News is a HIMSS Media publication.
Read more here:
AWS leader talks about technologies needed to take precision medicine to the next level - Healthcare IT News
Discover the theory of human decision-making using extensive experimentation and machine learning – Illinoisnewstoday.com
Discover a better theory
In recent years, the theory of human decision making has skyrocketed. However, these theories are often difficult to distinguish from each other and offer less improvement in explaining decision-making patterns than previous theories.Peterson et al. Leverage machine learning to evaluate classical decision theory, improve predictability, and generate new theories of decision making (see Perspectives by Bhatia and He). This method affects the generation of theory in other areas.
Science, Abe2629, this issue p. 1209abi7668, p. See also. 1150
Predicting and understanding how people make decisions is a long-standing goal in many areas, along with a quantitative model of human decision-making that informs both social science and engineering research. did. Show how large datasets can be used to accelerate progress towards this goal by enhancing machine learning algorithms that are constrained to generate interpretable psychological theories. .. Historical discoveries by conducting the largest experiments on risky choices to date and analyzing the results using gradient-based optimizations of differentiable decision theory implemented via artificial neural networks. A new, more accurate model of human decision-making in the form of summarizing, confirming that there is room for improvement of existing theories, and preserving insights from centuries of research.
Discover the theory of human decision-making using extensive experimentation and machine learning
Source link Discover the theory of human decision-making using extensive experimentation and machine learning
See the rest here:
Discover the theory of human decision-making using extensive experimentation and machine learning - Illinoisnewstoday.com