Category Archives: Machine Learning
Teaching tech teams every step of implementing a machine learning project – iTnews
Through a series of recent breakthroughs, even programmers who know close to nothing about deep learning technology can use simple, efficient tools to implement programs capable of learning from data.
If theyre lucky, they might find a book like OReillys Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow, which provides concrete examples to help learners gain an intuitive understanding of the concepts and tools for building intelligent systems. (See link below for free download of chapter 2, End-to-End Machine Learning Project.)
But if someone on your tech team wants to dive even deeper into MLor any new technology, for that matterwhere can they turn?
Learning platforms give tech teams a real advantage
The future of work is becoming increasingly remote. Certainly thats been accelerated by the COVID-19 pandemic, but many studies have shown that working from home, whether full time or on a hybrid schedule, is here to stay. But when youre working remotely, who do you turn to when youre stuck? Theres no cubicle mate to rely on. Slack messages arent always instantly returned. Where can remote workers go to get the solutions they need so they can get back to work fast?
Heres where learning and development (L&D) solutions can really shineand theyre only growing brighter as we look toward how well be working tomorrow. And accommodating remote work is just the latest in a long list of transitions L&D solutions have made to help employees over the decades.
The publisher of the book noted above, OReilly, is a learning company, and for over 40 years its worked to meet tech learners where they are. But where learners are has changed over time. Thats why OReilly has transformed itselfto its book publishing enterprise, it added live global conferences, and now it has consolidated all its services into one of the most comprehensive learning platforms for tech professionals. And that transformation has resonated across the industry. Today more than 60 percent of Fortune 100 companies count on the OReilly learning platform to train their tech teams.
How do learning platforms like OReilly work?
L&D solutionsespecially those that offer certificationsare a great way for employees to learn new skills and tools that they may not have encountered during their formal education. Thats particularly true for technology professionals; often by the time they graduate, the technologies they learned in school are already becoming obsolete.
But the potential for L&D goes beyond attaining greater knowledge or advancing in title or compensation. It can also help eliminate some of the headaches all tech teams experience while tackling their daily work. For remote employees in particular, its crucial to learn in the flow of workto quickly and easily find the answers they need to overcome obstacles and get back to the task at hand.
This leads to the most recent step in OReillys evolution to meet learners where they are (both figuratively, with what theyre learning, and literally, while theyre at home). It recently added AI-enabled capabilities to create OReilly Answers. Members simply ask a question, and the platform instantly scans thousands of titles to find the best answer that solves their tech problemsometimes down to a specific line of code. Imagine all the time saved from scouring pages upon pages of books! What once took hours now takes secondsa natural language processing engine is the new cubicle mate.
The OReilly learning platform also offers thousands of live online events and training courses, where teams can ask questions and get answers from industry experts. So now if they want to attend a tech conference, they can do so from homewithout all the travel costs. Another standout feature is OReillys interactive learning scenarios and sandboxes, where teams can access safe live development environments to practice with new technologies and tools before trying to put them to work in real-world situations.
Download chapter 2 of Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow for free
Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow is just one of over 60,000 titles that are available on the OReilly learning platform. In chapter 2, End-to-End Machine Learning Project, author Aurlien Gron walks you through every step of standing up an ML project, from seeing the big picture to launching and maintaining your system. And right now OReilly is offering a free download of the chapter so potential customers can see the high-quality content available on its learning platform. Its a highly recommended read.
More:
Teaching tech teams every step of implementing a machine learning project - iTnews
Machine-learning to speed up treatment of brain injury – Cosmos
A team of data scientists from the University of Pittsburgh School of Medicine in the US, and neurotrauma surgeons from the University of Pittsburgh Medical Centre, has developed the first automated brain scans and machine-learning techniques to inform outcomes for patients who have severe traumatic brain injuries.
The advanced machine-learning algorithm can analyse vast volumes of data from brain scans and relevant clinical data from patients. The researchers found that the algorithm was able to quickly and accurately produce a prognosis up to six months after injury. The sheer amount of data examined and the speed with which it is analysed is simply not possible for a human clinician, the researchers say.
More on machine learning in medicine: Are machine-learning tools the future of healthcare?
Publishing their results this week in Radiology, the scientists new predictive algorithm has been validated across two independent patient cohorts.
Co-senior author of the paper Shandong Wu, associate professor of radiology, bioengineering and biomedical informatics at University of Pittsburgh in the US, is an expert at using machine learning in medicine. The researchers used a hybrid model machine-learning framework using deep learning and traditional machine learning, processing CT imaging data and clinical non-imaging data for severe traumatic brain injury patient outcome prediction, he tells Cosmos.
Wu says the team used data from the University of Pittsburgh Medical Center (UPMC) and another 18 institutions from around the US. By using the machine learning model when the patient is admitted early in the emergency room, were able to build a model that can automatically predict favourable or unfavourable outcome or the mortality or the other recovery potential, he says.
Get an update of science stories delivered straight to your inbox.
We find our model maintains prediction performance, which shows our model is capturing some critical information to be able to provide that kind of prediction.
Co-senior author Dr David Okonkwo, a professor of neurological surgery at the University of Pittsburgh and a practising neurosurgeon, also spoke with Cosmos. After presenting the same data to a small group of neurosurgeons, Okonkwo says the machine learning model significantly outperformed human judgment and experience.
The success of the first model, based on specific data sets from within the first few hours of the injury, is extremely encouraging and telling us that were on the right path here to build tools that can complement human clinical judgment to make the best decisions for patients, says Okonkwo. But the researchers believe it can be made more powerful and accurate.
The first three-day window is very critical for better or for worse for patients with severe traumatic brain injuries. The most common reason for someone to die in the hospital after a traumatic brain injury is because of withdrawal of life-sustaining therapy, and this most commonly happens within the first 72 hours, Okonkwo says.
If we can build a model that is based off of that first three days worth of information, we think that we can put clinicians in a better place to identify the patients that have a chance at a meaningful recovery.
The study is one of many using machine learning in different areas of medicine, says Wu. There are tons of new leading research in the past couple of years, using all kinds of imaging or clinical data and machine learning or deep learning to address many other medical issues, diseases or conditions, he says.
Our study as on top of that, another strong study showing, you know, critical care and severe trauma and brain injury population, how our techniques or how deep learning can provide more information, or additional tools to help physicians like David here to provide improved care to patients. Okonkwo says machine-learning tools are intended not to replace human clinical or human judgment, but to complement human clinical decision making.
Read the original here:
Machine-learning to speed up treatment of brain injury - Cosmos
Deep learning is bridging the gap between the digital and the real world – VentureBeat
We are excited to bring Transform 2022 back in-person July 19 and virtually July 20 - 28. Join AI and data leaders for insightful talks and exciting networking opportunities. Register today!
Algorithms have always been at home in the digital world, where they are trained and developed in perfectly simulated environments. The current wave of deep learning facilitates AIs leap from the digital to the physical world. The applications are endless, from manufacturing to agriculture, but there are still hurdles to overcome.
To traditional AI specialists, deep learning (DL) is old hat. It got its breakthrough in 2012 when Alex Krizhevsky successfully deployed convolutional neural networks, the hallmark of deep learning technology, for the first time with his AlexNet algorithm. Its neural networks that have allowed computers to see, hear and speak. DL is the reason we can talk to our phones and dictate emails to our computers. Yet DL algorithms have always played their part in the safe simulated environment of the digital world. Pioneer AI researchers are working hard to introduce deep learning to our physical, three-dimensional world. Yep, the real world.
Deep learning could do much to improve your business, whether you are a car manufacturer, a chipmaker or a farmer. Although the technology has matured, the leap from the digital to the physical world has proven to be more challenging than many expected. This is why weve been talking about smart refrigerators doing our shopping for years, but no one actually has one yet. When algorithms leave their cozy digital nests and have to fend for themselves in three very real and raw dimensions there is more than one challenge to be overcome.
The first problem is accuracy. In the digital world, algorithms can get away with accuracies of around 80%. That doesnt quite cut it in the real world. If a tomato harvesting robot sees only 80% of all tomatoes, the grower will miss 20% of his turnover, says Albert van Breemen, a Dutch AI researcher who has developed DL algorithms for agriculture and horticulture in The Netherlands. His AI solutions include a robot that cuts leaves of cucumber plants, an asparagus harvesting robot and a model that predicts strawberry harvests. His company is also active in the medical manufacturing world, where his team created a model that optimizes the production of medical isotopes. My customers are used to 99.9% accuracy and they expect AI to do the same, Van Breemen says. Every percent of accuracy loss is going to cost them money.
To achieve the desired levels, AI models have to be retrained all the time, which requires a flow of constantly updated data. Data collection is both expensive and time-consuming, as all that data has to be annotated by humans. To solve that challenge Van Breemen has outfitted each of his robots with functionality that lets it know when it is performing either well or badly. When making mistakes the robots will upload only the specific data where they need to improve. That data is collected automatically across the entire robot fleet. So instead of receiving thousands of images, Van Breemens team only gets a hundred or so, that are then labeled and tagged and sent back to the robots for retraining. A few years ago everybody said that data is gold, he says. Now we see that data is actually a huge haystack hiding a nugget of gold. So the challenge is not just collecting lots of data, but the right kind of data.
His team has developed software that automates the retraining of new experiences. Their AI models can now train for new environments on their own, effectively cutting out the human from the loop. Theyve also found a way to automate the annotation process by training an AI model to do much of the annotation work for them. Van Breemen: Its somewhat paradoxical because you could argue that a model that can annotate photos is the same model I need for my application. But we train our annotation model with a much smaller data size than our goal model. The annotation model is less accurate and can still make mistakes, but its good enough to create new data points we can use to automate the annotation process.
The Dutch AI specialist sees a huge potential for deep learning in the manufacturing industry, where AI could be used for applications like defect detection and machine optimization. The global smart manufacturing industry is currently valued at 198 billion dollars and has a predicted growth rate of 11% until 2025. The Brainport region around the city of Eindhoven where Van Breemens company is headquartered is teeming with world-class manufacturing corporates, such as Philips and ASML. (Van Breemen has worked for both companies in the past.)
A second challenge of applying AI in the real world is the fact that physical environments are much more varied and complex than digital ones. A self-driving car that is trained in the US will not automatically work in Europe with its different traffic rules and signs. Van Breemen faced this challenge when he had to apply his DL model that cuts cucumber plant leaves to a different growers greenhouse. If this took place in the digital world I would just take the same model and train it with the data from the new grower, he says. But this particular grower operated his greenhouse with LED lighting, which gave all the cucumber images a bluish-purple glow our model didnt recognize. So we had to adapt the model to correct for this real-world deviation. There are all these unexpected things that happen when you take your models out of the digital world and apply them to the real world.
Van Breemen calls this the sim-to-real gap, the disparity between a predictable and unchanging simulated environment and the unpredictable, ever-changing physical reality. Andrew Ng, the renowned AI researcher from Stanford and cofounder of Google Brain who also seeks to apply deep learning to manufacturing, speaks of the proof of concept to production gap. Its one of the reasons why 75% of all AI projects in manufacturing fail to launch. According to Ng paying more attention to cleaning up your data set is one way to solve the problem. The traditional view in AI was to focus on building a good model and let the model deal with noise in the data. However, in manufacturing a data-centric view may be more useful, since the data set size is often small. Improving data will then immediately have an effect on improving the overall accuracy of the model.
Apart from cleaner data, another way to bridge the sim-to-real gap is by using cycleGAN, an image translation technique that connects two different domains, made popular by aging apps like FaceApp. Van Breemens team researched cycleGAN for its application in manufacturing environments. The team trained a model that optimized the movements of a robotic arm in a simulated environment, where three simulated cameras observed a simulated robotic arm picking up a simulated object. They then developed a DL algorithm based on cycleGAN that translated the images from the real world (three real cameras observing a real robotic arm picking up a real object) to a simulated image, which could then be used to retrain the simulated model. Van Breemen: A robotic arm has a lot of moving parts. Normally you would have to program all those movements beforehand. But if you give it a clearly described goal, such as picking up an object, it will now optimize the movements in the simulated world first. Through cycleGAN you can then use that optimization in the real world, which saves a lot of man-hours. Each separate factory using the same AI model to operate a robotic arm would have to train its own cycleGAN to tweak the generic model to suit its own specific real-world parameters.
The field of deep learning continues to grow and develop. Its new frontier is called reinforcement learning. This is where algorithms change from mere observers to decision-makers, giving robots instructions on how to work more efficiently. Standard DL algorithms are programmed by software engineers to perform a specific task, like moving a robotic arm to fold a box. A reinforcement algorithm could find out there are more efficient ways to fold boxes outside of their preprogrammed range.
It was reinforcement learning (RL) that made an AI system beat the worlds best Go player back in 2016. Now RL is also slowly making its way into manufacturing. The technology isnt mature enough to be deployed just yet, but according to the experts, this will only be a matter of time.
With the help of RL, Albert Van Breemen envisions optimizing an entire greenhouse. This is done by letting the AI system decide how the plants can grow in the most efficient way for the grower to maximize profit. The optimization process takes place in a simulated environment, where thousands of possible growth scenarios are tried out. The simulation plays around with different growth variables like temperature, humidity, lighting and fertilizer, and then chooses the scenario where the plants grow best. The winning scenario is then translated back to the three-dimensional world of a real greenhouse. The bottleneck is the sim-to-real gap, Van Breemen explains. But I really expect those problems to be solved in the next five to ten years.
As a trained psychologist I am fascinated by the transition AI is making from the digital to the physical world. It goes to show how complex our three-dimensional world really is and how much neurological and mechanical skill is needed for simple actions like cutting leaves or folding boxes. This transition is making us more aware of our own internal, brain-operated algorithms that help us navigate the world and which have taken millennia to develop. Itll be interesting to see how AI is going to compete with that. And if AI eventually catches up, Im sure my smart refrigerator will order champagne to celebrate.
Bert-Jan Woertman is the director of Mikrocentrum.
Welcome to the VentureBeat community!
DataDecisionMakers is where experts, including the technical people doing data work, can share data-related insights and innovation.
If you want to read about cutting-edge ideas and up-to-date information, best practices, and the future of data and data tech, join us at DataDecisionMakers.
You might even considercontributing an articleof your own!
Read More From DataDecisionMakers
See the article here:
Deep learning is bridging the gap between the digital and the real world - VentureBeat
Top 10 Artificial Intelligence Repositories on GitHub – Analytics Insight
Take a look at the top 10 artificial intelligence repositories on Github.GitHub
GitHub has become increasingly popular in no time. This is one of the most popular platforms for coders and developers to host and share codes in a cooperative and collaborative environment. GitHub boasts millions of repositories in various domains. In this article, we will throw light on the top 10 artificial intelligence repositories on GitHub. Have a look!
TensorFlow has gained wide recognition as an open-source framework for Machine learning and Artificial Intelligence. This GitHub repository was developed by Google Brain Team and contains various resources to learn. With the state-of-the-art models for computer vision, NLP, and recommendation systems, you are bound to generate highly accurate results on their datasets.
This is a lightweight TensorFlow-based network that is used for automatically learning high-quality models with the least expert interference. This AI repository on GitHub boasts easy usability, flexibility, speed, and a guarantee of learning.
BERT (Bidirectional Encoder Representations from Transformers) is the first unsupervised, deeply bidirectional system for pre-training NLP. Evidently enough, this AI repository contains TensorFlow code and pre-trained models for BERT, aimed at obtaining new state-of-the-art results on a significant number of NLP tasks.
This Artificial intelligence repository focuses majorly on data processing. However, a point that is worth a mention is that Airflow has the opinion that tasks should ideally be idempotent. In simple terms, the results of the task will be the same, and will not create duplicated data in a destination system
This is a beginner-level AI GitHub repository that evidently emphasises document similarity. The idea behind the document similarity application is to find the common topic discussed between the documents.
AI Learning is yet another most widely relied upon AI GitHub repository that consists of many lessons such as Machine Learning (ML), Deep Learning (DL), and Natural Language Processing, to name a few.
This GitHub repository is an exclusive Machine Learning sub-repository that contains various algorithms coded exclusively in Python. Here, you get codes on several regression techniques such as linear and polynomial regression. This repository finds immense application in predictive analysis for continuous data.
This AI repository on GitHub is widely recognized across the globe as it contains classification, regression, and clustering algorithms, as well as data-preparation and model-evaluation tools. Can it get any better than this?
This GitHub repository has an organized list of machine learning libraries, frameworks, and tools in almost all the languages available. All in all, Awesome Machine Learning promotes a collective development environment for Machine Learning.
spaCy is a library foradvanced Natural Language Processingin Python. spaCy is that one repository that is built on the very latest research and was designed from day one to be used in real products.
Share This ArticleDo the sharing thingy
About AuthorMore info about author
Follow this link:
Top 10 Artificial Intelligence Repositories on GitHub - Analytics Insight
Your AI can’t tell you it’s lying if it thinks it’s telling the truth. That’s a problem – The Register
Opinion Machine learning's abiding weakness is verification. Is your AI telling the truth? How can you tell?
This problem isn't unique to ML. It plagues chip design, bathroom scales, and prime ministers. Still, with so many new business models depending on AI's promise to bring the holy grail of scale to real-world data analysis, this lack of testability has new economic consequences.
The basic mechanisms of machine learning are sound, or at least statistically reliable. Within the parameters of its training data, an ML process will deliver what the underlying mathematics promise. If you understand the limits, you can trust it.
But what if there's a backdoor, a fraudulent tweak of that training data set which will trigger misbehavior? What if there's a particular quirk in someone's loan request submitted at exactly 00:45 on the 5th and the amount requested checksums to 7 that triggers automatic acceptance, regardless of risk?
Like an innocent assassin unaware they'd had a kill word implanted under hypnosis, your AI would behave impeccably until the bad guys wanted it otherwise.
Intuitively, we know that's a possibility. Now it has been shown mathematically that not only can this happen, researchers say, it's not theoretically detectable. An AI backdoor exploit engineered through training is not only just as much a problem as a traditionally coded backdoor, it's not amenable to inspection or version-on-version comparison or, indeed, anything. As far as the AI's concerned, everything is working perfectly, Harry Palmer could never confess to wanting to shoot JFK, he had no idea he did.
The mitigations suggested by researchers aren't very practical. Complete transparency of training data and process between AI company and client is a nice idea, except that the training data is the company's crown jewels and if they're fraudulent, how does it help?
At this point, we run into another much more general tech industry weakness, the idea that you can always engineer a singular solution to a particular problem. Pay the man, Janet, and let's go home. That doesn't work here; computer says no is one thing, mathematics says no quite another. If we carry on assuming that there'll be a fix akin to a patch, some new function that makes future AIs resistant to this class of fraud, we will be defrauded.
Conversely, the industry does genuinely advance once fundamental flaws are admitted and accepted, and the ecosystem itself changes in recognition.
AI has an ongoing history of not working as well as we thought, and it's not just this or that project. For example, an entire sub-industry has evolved to prove you are not a robot. Using its own trained robots to silently watch you as you move around online. If these machine monitors deem you too robotic, they spring a Voight-Kampff test on you in the guise of a Completely Automated Public Turing test to tell Computers and Humans Apart more widely known, and loathed, as a Captcha. You then have to pass a quiz designed to filter out automata. How undignified.
Do they work? It's still economically viable for the bad guys to carry on producing untold millions of programmatic fraudsters intent on deceiving the advertising industry, so that's a no on the false positives. And it's still common to be bounced from a login because your eyes aren't good enough, or the question too ambiguous, or the feature you relied on has been taken away. Not being able to prove you are not a robot doesn't get you shot by Harrison Ford, at least for now, but you may not be able to get into eBay.
The answer here is not to build a "better" AI and feed it with more and "better" surveillance signals. It's to find a different model to identify humans online, without endangering their privacy. That's not going to be a single solution invented by a company, that's an industry-wide adoption of new standards, new methods.
Likewise, you will never be able to buy a third-party AI that is testably pure of heart. To tell the truth, you'll never be able to build one yourself, at least not if you've got a big enough team or a corporate culture where internal fraud can happen. That's a team of two or more, and any workable corporate culture yet invented.
That's OK, once you stop looking for that particular unicorn. We can't theoretically verify non-trivial computing systems of any kind. When we have to use computers where failure is not an option, like flying aircraft or exploring space, we use multiple independent systems and majority voting.
If it seems that building a grand scheme on the back of the "perfect" black box works as badly as designing a human society on the model of the perfectly rational human, congratulations. Handling the complexities of real world data at real world scale means accepting that any system is fallible in ways that can't be patched or programmed out of. We're not at the point where AI engineering is edging into AI psychology, but it's coming.
Meanwhile, there's no need to give up on your AI-powered financial fraud detection. Buy three AIs from three different companies. Use them to check each other. If one goes wonky, use the other two until you can replace the first.
Can't afford three AIs? You don't have a workable business model. At least AI is very good at proving that.
AI Dynamics Will Employ Machine Learning to Triage TB Patients More Accurately, Quickly, Simply and Inexpensively Using Cough Sound Data, Bringing…
AI Dynamics
Selected by QB3 and UCSF for R2D2 TB Networks Scale Up Your TB Diagnostic Solution Program
BELLEVUE, Wash., April 26, 2022 (GLOBE NEWSWIRE) -- AI Dynamics, an organization founded on the belief that everyone should have access to the power of artificial intelligence (AI) to change the world, has been selected for the Rapid Research in Diagnostics Development for TB Networks (R2D2 TB Network) Scale Up Your TB Diagnostic Solution Program, hosted by QB3 and the UCSF Rosenman Institute. With 1.5 million deaths reported each year, Tuberculosis (TB) is the worldwide leading cause of death from a single infectious disease agent. The goal of the program is to harness machine learning technology for triaging TB using simple and affordable tests that can be performed on easy-to-collect samples such as cough sounds.
Currently, two weeks of cough sound data is widely used to determine who requires costly confirmatory testing, which delays the initiation of the treatment. AI Dynamics will build a proof-of-concept machine learning model to triage TB patients more accurately, quickly, simply and inexpensively using cough sounds, relieving patients from paying for unnecessary molecular and culture TB tests. Due to the prevalence of TB in under-resourced and remote locations, access to affordable early detection options is necessary to prevent disease transmissions and deaths in such countries.
At the core of AI Dynamics mission is providing equal access to the power of AI to everyone and we are committed to working with like-minded companies that recognize the positive impact innovative technology can have on the world, Rajeev Dutt, Founder and CEO of AI Dynamics said. The collaboration and accessible datasets that the R2D2 TB Network provides help to facilitate life-changing diagnostics for the most vulnerable populations.
The R2D2 TB Network offers a transparent and partner-engaged process for the identification, evaluation and advancement of promising TB diagnostics by providing experts and data and facilitating rigorous clinical study evaluation. AI Dynamics will build and validate a model using cough sounds collected from sites worldwide through the R2D2 TB Network.
Story continues
About AI Dynamics:
AI Dynamics aims to make artificial intelligence (AI) accessible to organizations of all sizes. The company's NeoPulse Framework is an intuitive development and management platform for AI, which enables companies to develop and implement deep neural networks and other machine learning models that can improve key performance metrics. The company's team brings decades of experience in the fields of machine learning and artificial intelligence from leading companies and research organizations. For more information, please visit aidynamics.com.
About The R2D2 TB Network:
The Rapid Research in Diagnostics Development for TB Network (R2D2 TB Network) brings together various TB experts with highly experienced clinical study sites in 10 countries. For further information, please visit their website at https://www.r2d2tbnetwork.org/.
Media Contact:
Justine GoodielUPRAISE Marketing + PR for AI Dynamicsaidynamics@upraisepr.com
AI, ML, & Cybersecurity: Here’s What FDA May Soon Be Asking – Design News
FDA has released a number of documents that could help clarify its expectations for artificial intelligence, machine learning, and cybersecurity. These include Artificial Intelligence/Machine Learning (AI/ML)-Based Software as a Medical Device (SaMD) Action Plan, published in January 2021; Good Machine Learning Practice for Medical Device Development: Guiding Principles, published in October 2021; and the just-released draft guidance, Cybersecurity in Medical Devices: Quality System Considerations and Content of Premarket Submissions.
Michelle Jump, vice president of security services for Medsec; and Yarmela Pavlovic, vice president of regulatory strategy for Medtronic, explored these documents during the IME West 2022 session on April 12, Product Development Innovation in an Age of Regulatory Uncertainty.
The AI/ML action plan provides a more tailored regulatory framework for AI/ML, explained Pavlovic. She referred to FDAs 2019 discussion paper, Proposed Regulatory Framework for Modifications to Artificial Intelligence/Machine Learning (AI/ML)-Based Software as a Medical Device (SaMD) - Discussion Paper and Request for Feedback, which laid out a total product lifecycle approach to AI/ML regulations with the understanding that AI/ML products can be iterated much more efficiently and quickly than a typical medical device implant product or something that isnt software based. This is because there is an opportunity to add additional data to training sets on which the products were originally formulated, she said.
Key to the total product lifecycle approach is a predetermined change control plan" (PCCP),which describes all the types of changes the manufacturer intends to make to the product in the foreseeable future and the protocols and success criteria and data sets they will use for evaluating the performance of these products, she explained.
In answering a question from the audience about PCCPs, Pavlovic explained that when you think about formulating a plan, you have to start with what the testing package you plan to give to FDA looks like. The level of detail of the PCCP will match the level of detail in the protocol and the test reports in your starting package. What you do in the future to validate a product flows from what you did in the first instance. And then the types of changes that you describe may warrant additional or differenttypes of testing, so be thoughtful about the key scientific questions that are posed by those changes and how can you layer on the right level of evidence to be comfortable.
The Medical Device Innovation Consortium has a digital health vertical with a workstream looking to build out PCCP examples, she added.
The AI/ML action plan also includes good machine learning practices (GMLP), such as having good hygiene around data sets and data management practices as well as processes to continue to learn about the product as it is used in commercial practice to make sure you are continuing to understand the performance of your product, explained Pavlovic.
Also,algorithmic bias and robustness is an area for further regulatory science development, she said. The goal would be to ensure the performance of the product is representative of the intended use population and that we dont have unintended consequences from the use of our products because of bias present due to the way they were trained or choices of data sets.
Real-world performance is the last element of the plan, she said, which is the idea that companies would monitor performance of the product in the wild, such as in clinical use. But it is complex to have info sharing agreements with customers, she said, so there may be the need to find other ways to gather such data.
Answering an audience question on FDAs stance toward AI, Pavlovic said that it is important for companies to make it clear to FDA that they are taking a rigorous approach to evaluation of a product and that there are processes in place to ensure continued performance once the product is on the market.
We have a responsibility to get these products right so we dont undermine forward progress. All of us in this room take that responsibility seriously. A couple poorly performing products will set us back, she added.
Jump, who has been in the digital health space for more than a decade, added that she has watched FDA become more comfortable with software. For instance, FDA wasnt initially comfortable with remote updates, but now they are saying, Why arent you doing remote software updates? she said.
But what FDA may not be comfortable with could be the reasoning behind a product, she said. If its new and innovative . . . get comfortable explaining what you are doing. When it comes to AI, for instance, FDA is concerned they see things going into a black box and coming out with an answer, she explained. They dont understand what could change that would result in an unexpected decision that clinicians are trusting.
Pavlovic and Jump also shared some initial feedback on the brand new April 2022 draft guidance on cybersecurity, which is much longer than the 2014 guidance. It may scare people . . . but it is a blessing in disguise, said Jump. Follow what this guidance says, because it is what they are going to ask for . . . they are already asking for [it].
For instance, just as medical device companies have to do risk management based on other regulation, the agency now expects threat modeling, she explained.
Get your comments in now, Jump added. I dont think theyre going to change a lot. Said Pavlovic: It is very consistent with what theyve been saying.
Jump pointed out a few new terms in the guidance, such as software product development framework (SPDF). You should have a secure design processwhich is a new name for an old concept. A good secure design is much more effective than constantly going in and patching things. FDA really wants to push better, secure design approaches.
She also clarified the scope of FDAs regulatory approach. You do not have to be a connected product to be applicable to cybersecurityif it has software or programmable logic, you are under scope of the guidance. She also added that if a company is submitting a change to hardware, FDA may still ask about cybersecurity.
Read the original post:
AI, ML, & Cybersecurity: Here's What FDA May Soon Be Asking - Design News
Are machine-learning tools the future of healthcare? – Cosmos
Terms like machine learning, artificial intelligence and deep learning have all become science buzzwords in recent years. But can these technologies be applied to saving lives?
The answer to that is a resounding yes. Future developments in health science may actually depend on integrating rapidly growing computing technologies and methods into medical practice.
Cosmos spoke with researchers from the University of Pittsburgh, in Pennsylvania, US, who have just published a paper in Radiology on the use of machine-learning techniques to analyse large data sets from brain trauma patients.
Co-lead author Shandong Wu, associate professor of radiology, is an authority on the use of machine learning in medicine. Machine-learning techniques have been around for several decades already, he explains. But it was in about 2012 that the so-called deep learning technique became mature. It attracted a lot of attention from the research field not only in medicine or healthcare, but in other domains, such as self-driving cars and robotics.
More on machine learning: Machine learning for cancer screening
So, what is deep learning? Its a kind of multi-layered, neural network-based model that is constantly mimicking how the human brain works to process a large set of data to learn or distill information, explains Wu.
The key to the increased maturity of machine-learning techniques in recent years is due to three interrelated developments, he says. These are the technical improvements in the algorithms of machine learning; the developments in the hardware being used, such as the improved graphical processing units; and the large volumes of digitised data readily available.
That data is key. Lots of it.
Get an update of science stories delivered straight to your inbox.
Machine-learning techniques use data to train the model to function better, and the more data the better. If you only have a small set of data, then you dont have a very good model, Wu explains. You may have very good questioning or good methodology, but youre not able to get a better model, because the model learns from lots of data.
Even though the available medical data is not as large as, say, social media data, there is still plenty to work with in the clinical domain.
Machine-learning models and algorithms can inform clinical decision-making, rapidly analysing massive amounts of data to identify patterns, says the papers other co-lead author, David Okonkwo.
Human beings can only process so much information. Machine learning permits orders of magnitude more information available than what an individual human can process, Okonkwo adds.
Okonkwo, a professor of neurological surgery, focuses on caring for patients with brain and spinal cord injuries, particularly those with traumatic brain injuries.
Our goal is to save lives, says Okonkwo. Machine-learning technologies will complement human experience and wisdom to maximise the decision-making for patients with serious injuries.
Even though today you dont see many examples, this will change the way that we practise medicine. We have very high hopes for machine learning and artificial intelligence to change the way that we treat many medical conditions from cancer, to making pregnancy safer, to solving the problems of COVID.
But important safeguards must be put in place. Okonkwo explains that institutions such as the US Food and Drugs Administration (FDA) must ensure that these new technologies are safe and effective before being used in real life-or-death scenarios.
Wu points out that the FDA has already approved about 150 artificial intelligence or machine learning-based tools. Tools need to be further developed or evaluated or used with physicians in the clinical settings to really examine their benefit for patient care, he says. The tools are not there to replace your physician, but to provide the tools and information to better inform physicians.
See original here:
Are machine-learning tools the future of healthcare? - Cosmos
Baseten nabs $20M to make it easier to build machine learning-based applications – TechCrunch
As the tech world inches a closer to the idea of artificial general intelligence, were seeing another interesting theme emerging in the ongoing democratization of AI: a wave of startups building tech to make AI technologies more accessible overall by a wider range of users and organizations.
Today, one of these, Baseten which is building tech to make it easier to incorporate machine learning into a business operations, production and processes without a need for specialized engineering knowledge is announcing $20 million in funding and the official launch of its tools.
These include a client API and a library of pre-trained models to deploy models built in TensorFlow, PyTorch or scikit-learn; the ability to build APIs to power your own applications; and the ability the create custom UIs for your applications based on drag-and-drop components.
The company has been operating in a closed, private beta for about a year and has amassed an interesting group of customers so far, including both Stanford and the University of Sydney, Cockroach Labs and Patreon, among others, who use it to, for example, help organizations with automated abuse detection (through content moderation) and fraud prevention.
The $20 million is being discussed publicly for the first time now to coincide with the commercial launch, and its in two tranches, with equally notable names among those backers.
The seed was co-led by Greylock and South Park Commons Fund, with participation also from the AI Fund, Caffeinated Capital and individuals including Greg Brockman, co-founder and CTO at general intelligence startup OpenAI; Dylan Field, co-founder and CEO of Figma; Mustafa Suleyman, co-founder of DeepMind; and DJ Patil, ex-chief scientist of the United States.
Greylock also led the Series A, with participation from South Park Commons, early Stripe exec Lachy Groom; Dev Ittycheria, CEO of MongoDB; Jay Simon, ex-president of Atlassian, now at Bond; Jean-Denis Greze, CTO of Plaid; and Cristina Cordova, another former Stripe exec.
Tuhin Srivastava, Basetens co-founder and CEO, said in an interview that the funding will be used in part to bring on more technical and product people, and to ramp up its marketing and business development.
The issue that Baseten has identified and is trying to solve is one that is critical in the evolution of AI: Machine learning tools are becoming ever more ubiquitous and utilized, thanks to cheaper computing power, better access to training models and a growing understanding of how and where they can be used. But one area where developers still need to make a major leap, and businesses still need to make big investments, is when it comes to actually adopting and integrating machine learning: There remains a wide body of technical knowledge that developers and data scientists need to actually integrate machine learning into their work.
We were born out of the idea that machine learning will have a massive impact on the world, but its still difficult to extract value from machine learning models, Srivastava said. Difficult, because developers and data scientists need to have specific knowledge of how to handle machine learning ops, as well as technical expertise to manage production at the back end and the front end, he said. This is one reason why machine learning programs in businesses often actually have very little success: It takes too much effort to get them into production.
This is something that Srivastava and his co-founders Amir Haghighat (CTO) and Philip Howes (chief scientist) experienced firsthand when they worked together at Gumroad. Haghighat, who was head of engineering, and Srivastava and Howes, who were data scientists, wanted to use machine learning at the payments company to help with fraud detection and content moderation and realized that they needed to pick up a lot of extra full-stack engineering skills or hire specialists to build and integrate that machine learning along with all of the tooling needed to run it (e.g., notifications and integrating that data into other tools to action).
They built the systems still in use, and screening hundreds of millions of dollars of transactions but also picked up an idea in the process: Others surely were facing the same issues they did, so why not work on a set of tools to help all of them and take away some of that work?
Today, the main customers of Baseten a reference to base ten blocks, often used to help younger students learn the basics of mathematics (It humanizes the numbers system, and we wanted to make machine learning less abstract, too, said the CEO) are developers and data scientists who are potentially adopting other machine learning models, or even building their own but lack the skills to practically incorporate them into their own production flows. There, Baseten is part of a bigger group of companies that appear to be emerging building MLops solutions full sets of tools to make machine learning more accessible and usable by those working in DevOps and product. These include Databricks, Clear, Gathr and more. The idea here is to give tools to technical people to give them more power and more time to work on other tasks.
Baseten gets the process of tool-building out of the way so we can focus on our key skills: modeling, measurement and problem-solving, said Nikhil Harithras, senior machine learning engineer at Patreon, in a statement. Patreon is using Baseten to help run an image classification system, used to find content that violates its community guidelines.
Over time, there a logical step that Baseten could make, continuing on its democratization trajectory: considering how to build tools for non-technical audiences, too an interesting idea in light of the many no-code and low-code products that are being rolled out to give them more power to build their own data science applications, too.
Non-technical audiences are not something we focus on today, but that is the evolution, Srivastava said. The highest level goal is to accelerate the impact of machine learning.
More:
Baseten nabs $20M to make it easier to build machine learning-based applications - TechCrunch
Which Animal Viruses Could Infect People? Computers Are Racing to Find Out. – The New York Times
Colin Carlson, a biologist at Georgetown University, has started to worry about mousepox.
The virus, discovered in 1930, spreads among mice, killing them with ruthless efficiency. But scientists have never considered it a potential threat to humans. Now Dr. Carlson, his colleagues and their computers arent so sure.
Using a technique known as machine learning, the researchers have spent the past few years programming computers to teach themselves about viruses that can infect human cells. The computers have combed through vast amounts of information about the biology and ecology of the animal hosts of those viruses, as well as the genomes and other features of the viruses themselves. Over time, the computers came to recognize certain factors that would predict whether a virus has the potential to spill over into humans.
Once the computers proved their mettle on viruses that scientists had already studied intensely, Dr. Carlson and his colleagues deployed them on the unknown, ultimately producing a short list of animal viruses with the potential to jump the species barrier and cause human outbreaks.
In the latest runs, the algorithms unexpectedly put the mousepox virus in the top ranks of risky pathogens.
Every time we run this model, it comes up super high, Dr. Carlson said.
Puzzled, Dr. Carlson and his colleagues rooted around in the scientific literature. They came across documentation of a long-forgotten outbreak in 1987 in rural China. Schoolchildren came down with an infection that caused sore throats and inflammation in their hands and feet.
Years later, a team of scientists ran tests on throat swabs that had been collected during the outbreak and put into storage. These samples, as the group reported in 2012, contained mousepox DNA. But their study garnered little notice, and a decade later mousepox is still not considered a threat to humans.
If the computer programmed by Dr. Carlson and his colleagues is right, the virus deserves a new look.
Its just crazy that this was lost in the vast pile of stuff that public health has to sift through, he said. This actually changes the way that we think about this virus.
Scientists have identified about 250 human diseases that arose when an animal virus jumped the species barrier. H.I.V. jumped from chimpanzees, for example, and the new coronavirus originated in bats.
Ideally, scientists would like to recognize the next spillover virus before it has started infecting people. But there are far too many animal viruses for virologists to study. Scientists have identified more than 1,000 viruses in mammals, but that is most likely a tiny fraction of the true number. Some researchers suspect mammals carry tens of thousands of viruses, while others put the number in the hundreds of thousands.
To identify potential new spillovers, researchers like Dr. Carlson are using computers to spot hidden patterns in scientific data. The machines can zero in on viruses that may be particularly likely to give rise to a human disease, for example, and can also predict which animals are most likely to harbor dangerous viruses we dont yet know about.
It feels like you have a new set of eyes, said Barbara Han, a disease ecologist at the Cary Institute of Ecosystem Studies in Millbrook, N.Y., who collaborates with Dr. Carlson. You just cant see in as many dimensions as the model can.
Dr. Han first came across machine learning in 2010. Computer scientists had been developing the technique for decades, and were starting to build powerful tools with it. These days, machine learning enables computers to spot fraudulent credit charges and recognize peoples faces.
But few researchers had applied machine learning to diseases. Dr. Han wondered if she could use it to answer open questions, such as why less than 10 percent of rodent species harbor pathogens known to infect humans.
She fed a computer information about various rodent species from an online database everything from their age at weaning to their population density. The computer then looked for features of the rodents known to harbor high numbers of species-jumping pathogens.
Once the computer created a model, she tested it against another group of rodent species, seeing how well it could guess which ones were laden with disease-causing agents. Eventually, the computers model reached an accuracy of 90 percent.
Then Dr. Han turned to rodents that have yet to be examined for spillover pathogens and put together a list of high-priority species. Dr. Han and her colleagues predicted that species such as the montane vole and Northern grasshopper mouse of western North America would be particularly likely to carry worrisome pathogens.
Of all the traits Dr. Han and her colleagues provided to their computer, the one that mattered most was the life span of the rodents. Species that die young turn out to carry more pathogens, perhaps because evolution put more of their resources into reproducing than in building a strong immune system.
These results involved years of painstaking research in which Dr. Han and her colleagues combed through ecological databases and scientific studies looking for useful data. More recently, researchers have sped this work up by building databases expressly designed to teach computers about viruses and their hosts.
In March, for example, Dr. Carlson and his colleagues unveiled an open-access database called VIRION, which has amassed half a million pieces of information about 9,521 viruses and their 3,692 animal hosts and is still growing.
Databases like VIRION are now making it possible to ask more focused questions about new pandemics. When the Covid pandemic struck, it soon became clear that it was caused by a new virus called SARS-CoV-2. Dr. Carlson, Dr. Han and their colleagues created programs to identify the animals most likely to harbor relatives of the new coronavirus.
SARS-CoV-2 belongs to a group of species called betacoronaviruses, which also includes the viruses that caused the SARS and MERS epidemics among humans. For the most part, betacoronaviruses infect bats. When SARS-CoV-2 was discovered in January 2020, 79 species of bats were known to carry them.
But scientists have not systematically searched all 1,447 species of bats for betacoronaviruses, and such a project would take many years to complete.
By feeding biological data about the various types of bats their diet, the length of their wings, and so on into their computer, Dr. Carlson, Dr. Han and their colleagues created a model that could offer predictions about the bats most likely to harbor betacoronaviruses. They found over 300 species that fit the bill.
Since that prediction in 2020, researchers have indeed found betacoronaviruses in 47 species of bats all of which were on the prediction lists produced by some of the computer models they had created for their study.
Daniel Becker, a disease ecologist at the University of Oklahoma who also worked on the betacoronavirus study, said it was striking the way simple features such as body size could lead to powerful predictions about viruses. A lot of it is the low-hanging fruit of comparative biology, he said.
Dr. Becker is now following up from his own backyard on the list of potential betacoronavirus hosts. It turns out that some bats in Oklahoma are predicted to harbor them.
If Dr. Becker does find a backyard betacoronavirus, he wont be in a position to say immediately that it is an imminent threat to humans. Scientists would first have to carry out painstaking experiments to judge the risk.
Dr. Pranav Pandit, an epidemiologist at the University of California at Davis, cautions that these models are very much a work in progress. When tested on well-studied viruses, they do substantially better than random chance, but could do better.
Its not at a stage where we can just take those results and create an alert to start telling the world, This is a zoonotic virus, he said.
Nardus Mollentze, a computational virologist at the University of Glasgow, and his colleagues have pioneered a method that could markedly increase the accuracy of the models. Rather than looking at a viruss hosts, their models look at its genes. A computer can be taught to recognize subtle features in the genes of viruses that can infect humans.
In their first report on this technique, Dr. Mollentze and his colleagues developed a model that could correctly recognize human-infecting viruses more than 70 percent of the time. Dr. Mollentze cant yet say why his gene-based model worked, but he has some ideas. Our cells can recognize foreign genes and send out an alarm to the immune system. Viruses that can infect our cells may have the ability to mimic our own DNA as a kind of viral camouflage.
When they applied the model to animal viruses, they came up with a list of 272 species at high risk of spilling over. Thats too many for virologists to study in any depth.
You can only work on so many viruses, said Emmie de Wit, a virologist at Rocky Mountain Laboratories in Hamilton, Mont., who oversees research on the new coronavirus, influenza and other viruses. On our end, we would really need to narrow it down.
Dr. Mollentze acknowledged that he and his colleagues need to find a way to pinpoint the worst of the worst among animal viruses. This is only a start, he said.
To follow up on his initial study, Dr. Mollentze is working with Dr. Carlson and his colleagues to merge data about the genes of viruses with data related to the biology and ecology of their hosts. The researchers are getting some promising results from this approach, including the tantalizing mousepox lead.
Other kinds of data may make the predictions even better. One of the most important features of a virus, for example, is the coating of sugar molecules on its surface. Different viruses end up with different patterns of sugar molecules, and that arrangement can have a huge impact on their success. Some viruses can use this molecular frosting to hide from their hosts immune system. In other cases, the virus can use its sugar molecules to latch on to new cells, triggering a new infection.
This month, Dr. Carlson and his colleagues posted a commentary online asserting that machine learning may gain a lot of insights from the sugar coating of viruses and their hosts. Scientists have already gathered a lot of that knowledge, but it has yet to be put into a form that computers can learn from.
My gut sense is that we know a lot more than we think, Dr. Carlson said.
Dr. de Wit said that machine learning models could some day guide virologists like herself to study certain animal viruses. Theres definitely a great benefit thats going to come from this, she said.
But she noted that the models so far have focused mainly on a pathogens potential for infecting human cells. Before causing a new human disease, a virus also has to spread from one person to another and cause serious symptoms along the way. Shes waiting for a new generation of machine learning models that can make those predictions, too.
What we really want to know is not necessarily which viruses can infect humans, but which viruses can cause an outbreak, she said. So thats really the next step that we need to figure out.
Read more from the original source:
Which Animal Viruses Could Infect People? Computers Are Racing to Find Out. - The New York Times