Category Archives: Machine Learning
How machine learning helps the New York Times power its paywall – VentureBeat
Were you unable to attend Transform 2022? Check out all of the summit sessions in our on-demand library now! Watch here.
Every organization applying artificial intelligence (AI) and machine learning (ML) to their business is looking to use these powerful technologies to tackle thorny problems. For the New York Times, one of the biggest challenges is striking a balance between meeting its latest target of 15 million digital subscribers by 2027 while also getting more people to read articles online.
These days, the multimedia giant is digging into that complex cause-and-effect relationship using a causal machine learning model, called the Dynamic Meter, which is all about making its paywall smarter. According to Chris Wiggins, chief data scientist at the New York Times, for the past three or four years the company has worked to understand their user journey and the workings of the paywall.
Back in 2011, when the Times began focusing on digital subscriptions, metered access was designed so that non-subscribers could read the same fixed number of articles every month before hitting a paywall requiring a subscription. That allowed the company to gain subscribers while also allowing readers to explore a range of offerings before committing to a subscription.
Now, however, the Dynamic Meter can set personalized meter limits. By powering the model with data-driven user insights, the causal machine learning model can be prescriptive, determining the right number of free articles each user should get so they seem interested enough in the New York Times to subscribe in order to continue reading.
MetaBeat 2022
MetaBeat will bring together thought leaders to give guidance on how metaverse technology will transform the way all industries communicate and do business on October 4 in San Francisco, CA.
According to a blog post written by Rohit Supekar, a data scientist on the New York Times algorithmic targeting team, at the top of the sites subscription funnel are unregistered users. At a specific meter limit, they are shown a registration wall that blocks access and asks them to create an account. This allows them access to more free content, and a registration ID allows the company to better understand their activity. Once registered users reach another meter limit, they are served a paywall with a subscription offer. The Dynamic Meter model learns from all of this registered user data and determines the appropriate meter limit to optimize for specific key performance indicators (KPIs).
The idea, said Wiggins, is to form a long-term relationship with readers. Its a much slower problem [to solve], in which people engage over the span of weeks or months, he said.
The most difficult challenge in building the causal machine learning model was to set up the robust data pipeline that helps the algorithmic targeting team understand activity for over 130 million registered users on the New York Times site, said Supekar.
The key technical advancement powering the Dynamic Meter is around causal AI, a machine learning method where models are built which can predict not just will happen, but what would have happened.
Were really trying to understand the cause and effect, he explained.
If a particular user is given a different number of free articles, what would be the likelihood that they would subscribe or the likelihood that they would read a certain number of articles? This is a complicated question, he explained, because in reality, they can only observe one of these outcomes.
If we give somebody 100 free articles, we have to guess what would have happened if they were given 50 articles, he said. These sorts of questions fall in the realm of causal AI.
Supekars blog post explained that its clear how the causal machine learning model works by performing a randomized control trial, where certain groups of people are given different numbers of free articles and the model can learn based on this data. As the meter limit for registered users increases, the engagement measured by the average number of page views gets larger. But it also leads to a reduction in subscription conversions because fewer users encounter the paywall. The Dynamic Meter has to both optimize for and balance a trade-off between conversion engagement.
For a specific user who got 100 free articles, we can determine what would have happened if they got 50 because we can compare them with other registered users who were given 50 articles, said Supekar. This is an example of why causal AI has become popular: There are a lot of business decisions, which have a lot of revenue impact in our case, where we would like to understand the relationship between what happened and what would have happened, he explained. Thats where causal AI has really picked up steam.
Wiggins added that as more and more organizations bring AI into their businesses for automated decision-making, they really want to understand what is going on, at all angles.
Its different from machine learning in the service of insights, where you do a classification problem once and maybe you study that as a model, but you dont actually put the ML into production to make decisions for you, he said. Instead, for a business that wants AI to really make decisions, they want to have an understanding of whats going on. You dont want it to be a blackbox model, he pointed out.
Supekar added that his team is conscious of algorithmic ethics when it comes to the Dynamic Meter model. Our exclusive first-party data is only about the engagement people have with the Times content, and we dont include any demographic or psychographic features, he said.
As for the future of the New York Times paywall, Supekar said he is excited about exploring the science about the negative aspects of introducing paywalls in the media business.
We do know if you show paywalls we get a lot of subscribers, but we are also interested in knowing how a paywall affects some readers habits and the likelihood they would want to return in the future, even months or years down the line, he said. We want to maintain a healthy audience so they can potentially become subscribers, but also serve our product mission to increase readership.
The subscription business model has these kinds of inherent challenges, added Wiggins.
You dont have those challenges if your business model is about clicks, he said. We think about how our design choices now impact whether someone will continue to be a subscriber in three months, or three years. Its a complex science.
VentureBeat's mission is to be a digital town square for technical decision-makers to gain knowledge about transformative enterprise technology and transact. Discover our Briefings.
Here is the original post:
How machine learning helps the New York Times power its paywall - VentureBeat
The Democratization of Machine Learning: Heres how Applied Machine Learning Days (AMLD) Africa impacted (almost) the whole continent – African…
AMLD Africa (https://bit.ly/3AvIECg), a free Machine Learning conference, makes it possible for anyone in Africa to learn about AI with world-class speakers and entities.
15.7 trillion US$. This is the expected contribution of Artificial Intelligence to the global economy by 2030 according to PwCs report Sizing the Price (https://pwc.to/3R8JFY4). Heres how Applied Machine Learning Days (AMLD) Africa is offering free access to high-quality education around AI to democratize AI in Africa, and ultimately, contribute to making sure Africa has a fair share of the Price.
The more African can learn, grasp, and be inspired by AI, the more existing projects (or new ones) will leverage data to create a social, economic, and cultural value in local environments. It is with the same ambitious vision that AMLD Africas second edition will present AI through an African Lens from the 3rd to the 5th of November at the prestigious Mohammed VI Polytechnic University (UM6P) in Ben Guerir, Morocco.
AMLD Africa is a 3-day conference that consists of both inspiring talks and instructive workshops. Speakers will have an opportunity to inspire African talents, teach those who would like to improve their technical skills and strengthen the African Data Science community such as Zindi. Consequently, AMLD Africas conference will embody its motto: Democratizing Machine Learning in Africa.
For their first edition, AMLD Africa was able to present a truly comprehensive platform that included academics from Stanford, the University of the Western Cape and EPFL, corporates from IBM, Google, entrepreneurs and even the Assistant Director of the UNESCO. 3000 participants from all over the continent (50 African countries) were able to contemplate a true picture of AI, a picture where every entity was able to add its colour to make the whole as accurate as possible. If you too are a motivated and passionate AI enthusiast, you can embellish this years painting by clickinghere(https://bit.ly/3R60MKl) to apply for a talk or a workshop.
When asked on how to retain talent in the african continent: It is a matter of identifying important problems, and having the opportunity and tools to solve them. Furthermore, the increase of start-ups creates a dynamic and go-ahead environment for machine learning engineers and researches. If we manage to create such an ecosystem, we will be able to bring change, through solving long lasting problems. A future of Shared AI Knowledge, Opening Keynote, Moustapha Cisse Head of the Google AI Center in Accra, Ghana.
AMLD Africa not only includes both entities and individuals, but also has a cross-industry approach. Since AI is impacting every sector finance, national security, healthcare, etc. of private and public life, AMLD Africa presents the talks within tracks: Healthcare, agriculture and even entertainment for example. Whether it is Detecting cervical cancer with a smartphone-based solution, Measuring and optimizing agricultural production using aerial imagery or even music generated by AI (that you can listen to in the video below), the talks tackle actual challenges and conveyed the idea that technology is not a goal itself, but rather a tool for the minds and sometimes the ears.
AI has a lot of ethical, social, and economic challenges, but these challenges come with an undeniable opportunity to leapfrog the technological infrastructure associated with the Third Industrial Revolution. It is only when we give the young, motivated and passionate talents easy access to technology, that more initiatives will leverage data, and that AI could benefit daily life. If you are wondering how that can do, then AMLD Africa is the conference to attend.
Distributed by APO Group on behalf of AMLD Africa.
Contact:Mohamed Ali Dhraief[emailprotected]
This Press Release has been issued by APO. The content is not monitored by the editorial team of African Business and not of the content has been checked or validated by our editorial teams, proof readers or fact checkers. The issuer is solely responsible for the content of this announcement.
Here is the original post:
The Democratization of Machine Learning: Heres how Applied Machine Learning Days (AMLD) Africa impacted (almost) the whole continent - African...
GBT is Implementing Machine Learning Driven, Pattern Matching Technology for its Epsilon, Mi-crochip Reliability Verification and Correction EDA Tool…
GBT Technologies Inc.
SAN DIEGO, Sept. 01, 2022 (GLOBE NEWSWIRE) -- GBT Technologies Inc. (OTC PINK:GTCH) ("GBT or the Company), is implementing a machine learning driven, pattern matching technology within its Epsilon, microchips reliability verification and correction Electronic Design Automation (EDA) tool. Design rules are getting increasingly complex with each new process node and design firms are facing new challenges in the physical verification domain. One of the major areas that are affected by the process physics, is reliability Verification (RV). Microchips are major components nearly in every major electronics application. Civil, military and space exploration industries require reliable operations for many years, and in severe environments. High performance computing systems require advanced processing with high reliability to ensure the consistency and accuracy of the processed data. Complex integrated circuits are in the heart of these systems and need to function with high level of dependability. Particularly in the fields of medicine, aviation, transportation, data storage and industrial instrumentation, microchips reliability factor is crucial. GBT is implementing new machine learning driven, pattern matching techniques within its Epsilon system with the goal of addressing the advanced semiconductors physics, ensuring high level of reliability, optimal power consumption and high performance. As Epsilon analyzes the layout of an integrated circuit (IC), it identifies reliability weak spots, which are specific regions of an ICs layout, and learns their patterns. As the tool continues analyzing the layout it records problematic zones taking into account the patterns orientations and placements. In addition, it is designed to understand small variations in dimensions of the pattern, as specified by the designer or an automatic synthesis tool. As the weak spots are identified, the tool will take appropriate action to modify and correct them. A deep learning mechanism will be performing the data analysis, identification, categorization, and reasoning while executing an automatic correction. The Machine Learning will understand the patterns and record them in an internal library for future use. Epsilons pattern matching technology will be analyzing the chips data according to a set of predefined and learned-from-experience rules. Its cognitive capabilities will make it self-adjust to newest nodes with new constraints and challenges, with the goal of providing quick and reliable verification and correction of an IC layout.
Story continues
The Company released a video which explain the potential functions of the Epsilon tool: https://youtu.be/Mz4IOGRHeqw
"The ability to analyze and address advanced ICs reliability parameters is necessary to mitigate risk of system degradation, overheating, and possible malfunction. It can affect microchips performance, power consumption, data storage and retrieval, heat and an early failure which may be critical in vital electronic systems. Epsilon analyzes a microchip data for reliability, power and electrothermal characteristics, and performs auto-correction in case violations found. We are now implementing an intelligent technology for Epsilon with the goal of utilizing pattern matching algorithms to formulate a smart detection of reliability issues within integrated circuits layout. The new techniques will analyze and learn weak spots within microchips data, predicting failure models that are based on the process physics and electrical constraints knowledge. It will take into consideration each devices function, connectivity attributes, electrical currents information, electrothermal factors and more to determine problematic spots and perform auto-correction. Particularly for FinFet and GAA FET (Gate All Around FET) technologies, a devices functionality is developed with major reliability considerations ensuring power management efficiency, optimal thermal analysis aiming for long, reliable life span. Using smart pattern matching methods, we plan to improve reliability analysis, achieving consistency and accuracy across designs within advanced manufacturing processes. As dimensions of processes shrink, ICs layout features become much more complex to analyze for electrical phenomenon. To provide an intelligent answer for these complexities, we are implementing deep learning-based pattern matching technology with the goal of ensuring efficient, green microchips power consumption, higher performance, optimized thermal distribution, and ultimately superior reliability stated Danny Rittman, the Companys CTO.
There is no guarantee that the Company will be successful in researching, developing or implementing this system. In order to successfully implement this concept, the Company will need to raise adequate capital to support its research and, if successfully researched and fully developed, the Company would need to enter into a strategic relationship with a third party that has experience in manufacturing, selling and distributing this product. There is no guarantee that the Company will be successful in any or all of these critical steps.
About Us
GBT Technologies, Inc. (OTC PINK: GTCH) (GBT) (http://gbtti.com) is a development stage company which considers itself a native of Internet of Things (IoT), Artificial Intelligence (AI) and Enabled Mobile Technology Platforms used to increase IC performance. GBT has assembled a team with extensive technology expertise and is building an intellectual property portfolio consisting of many patents. GBTs mission, to license the technology and IP to synergetic partners in the areas of hardware and software. Once commercialized, it is GBTs goal to have a suite of products including smart microchips, AI, encryption, Blockchain, IC design, mobile security applications, database management protocols, with tracking and supporting cloud software (without the need for GPS). GBT envisions this system as a creation of a global mesh network using advanced nodes and super performing new generation IC technology. The core of the system will be its advanced microchip technology; technology that can be installed in any mobile or fixed device worldwide. GBTs vision is to produce this system as a low cost, secure, private-mesh-network between all enabled devices. Thus, providing shared processing, advanced mobile database management and sharing while using these enhanced mobile features as an alternative to traditional carrier services.
Forward-Looking Statements
Certain statements contained in this press release may constitute "forward-looking statements". Forward-looking statements provide current expectations of future events based on certain assumptions and include any statement that does not directly relate to any historical or current fact. Actual results may differ materially from those indicated by such forward-looking statements because of various important factors as disclosed in our filings with the Securities and Exchange Commission located at their website (http://www.sec.gov). In addition to these factors, actual future performance, outcomes, and results may differ materially because of more general factors including (without limitation) general industry and market conditions and growth rates, economic conditions, governmental and public policy changes, the Companys ability to raise capital on acceptable terms, if at all, the Companys successful development of its products and the integration into its existing products and the commercial acceptance of the Companys products. The forward-looking statements included in this press release represent the Company's views as of the date of this press release and these views could change. However, while the Company may elect to update these forward-looking statements at some point in the future, the Company specifically disclaims any obligation to do so. These forward-looking statements should not be relied upon as representing the Company's views as of any date subsequent to the date of the press release.
Contact:
Dr. Danny Rittman, CTO press@gopherprotocol.com
What’s the Difference Between Vertical Farming and Machine Learning? – Electronic Design
What youll learn
Sometimes inspiration comes in the oddest ways. I like to watch CBS News Sunday Morning because of the variety of stories they air. Recently, they did one on Vertical Farming - A New Form of Agriculture (see video below).
CBS News Sunday Morning recently did a piece on vertical farming that spawned this article.
For those who didnt watch the video, vertical farming is essentially a method of indoor farming using hydroponics. Hydroponics isnt new; its a subset of hydroculture where crops are grown without soil. Instead, the plants grow in a mineral-enriched water. This can be done in conjunction with sunlight but typically an artificial light source is used.
The approach is useful in areas that dont provide enough light, or at times or in locations where the temperature or conditions outside would not be conducive for growing plants.
Vertical farming is hydroponics taken to the extreme, with stacks upon stacks of trays with plants under an array of lights. These days, the lights typically are LEDs because of their efficiency and the ability to generate the type of light most useful for plant growth. Automation can be used to streamline planting, support, and harvesting.
A building can house a vertical farm anywhere in the world, including in the middle of a city. Though lots of water is required, its recycled, making it more efficient than other forms of agriculture.
Like many technologies, the opportunities are great if you ignore the details. Thats where my usual contrary nature came into play, though, since I followed up my initial interest by looking for limitations or problems related to vertical farming. Of course, I found quite a few and then noticed that many of the general issues applied to another topic I cover a lotmachine learning/artificial intelligence (ML/AI).
If you made it this far, you know how Im looking at the difference between machine learning and vertical farming. They obviously have no relationship in terms of their technology and implementation, but they do have much in common when one looks at the potential problems and solutions related to those technologies.
As electronic system designers and developers, we constantly deal with potential solutions and their tradeoffs. Machine learning is one of those generic categories that has proven useful in many instances. However, one must be wary of the issues underlying those flashy approaches.
Vertical farming, like machine learning, is something one can dabble in. To be successful, though, it helps to have an expert or at least someone who can quickly gain that experience. This tends to be the case with new and renewed technologies in general. I suspect significantly more ML experts are available these days for a number of reasons like the cost of hardware, but the demand remains high.
Vertical farming uses a good bit of computer automation. The choice of plants, fertilizers, and other aspects of hydropic farming are critical to the success of the farm. Then theres the maintenance aspect. ML-based solutions are one way of reducing the expertise or time required by the staff to support the system.
ML programmers and developers also are able to obtain easier-to-use tools, thereby reducing the amount of expertise and training required to take advantage of ML solutions. These tools often incorporate their own ML models, which are different than those being generated.
Hydroponics works well for many plants, but unfortunately for multiple others, thats not the case. For example, crops like microgreens work well. However, a cherry or apple tree often struggles with this treatment.
ML suffers from the same problem in that its not applicable to all computational chores. But, unlike vertical farms, ML applications and solutions are more diverse. The challenge for developers comes down to understanding where ML is and isnt applicable. Trying to force-fit a machine-learning model to handle a particular problem can result in a solution that provides poor results at high cost.
Vertical farms require power for lighting and to move liquid. ML applications tend to do lots of computation and thus require a good deal of power compared to other computational requirements. One big difference between the two is that ML solutions are scalable and hardware tradeoffs can be significant.
For example, ML hardware can improve performance thats orders of magnitude better than software solutions while reducing power requirements. Likewise, even software-only solutions may be efficient enough to do useful work even while using little power, simply because developers have made the ML models work within the limitations of their design. Vertical farms do not have this flexibility.
Large vertical farms do require a major investment, and theyre not cheap to run due to their scale. The same is true for cloud-based ML solutions utilizing the latest in disaggregated cloud-computing centers. Such data centers are leveraging technologies like SmartNIC and smart storage to use ML models closer to communication and storage than was possible in the past.
The big difference with vertical farming versus ML is scalability. Its now practical for multiple ML models to be running in a smartwatch with a dozen sensors. But that doesnt compare to dealing with agriculture that must scale with the rest of the physical world requirements, such as the plants themselves.
Still, these days, ML does require a significant investment with respect to development and developing the experience to adequately apply ML. Software and hardware vendors have been working to lower both the startup and long-term development costs, which has been further augmented by the plethora of free software tools and low-cost hardware thats now generally available.
Cut the power on a vertical farm and things come to a grinding halt rather quickly, although its not like having an airplane lose power at 10,000 feet. Still, plants do need sustenance and light, though theyre accustomed to changes over time. Nonetheless, responding to failures within the system is important to the systems long-term usefulness.
ML applications tend to require electricity to run, but that tends to be true of the entire system. A more subtle problem with ML applications is the source of input, which is typically sensors such as cameras, temperature sensors, etc. Determining whether the input data is accurate can be challenging; in many cases, designers simply assume that this information is accurate. Applications such as self-driving cars often use redundant and alternative inputs to provide a more robust set of inputs.
Vertical-farming technology continues to change and become more refined, but its still maturing. The same is true for machine learning, though the comparison is like something between a penny bank and Fort Knox. There are simply more ML solutions, many of which are very mature with millions of practical applications.
That said, ML technologies and applications are so varied, and the rate of change so large, that keeping up with whats availablelet alone how things work in detailcan be overwhelming.
Vertical farming is benefiting from advances in technology from robotics to sensors to ML. The ability to track plant growth, germination, and detecting pests are just a few tasks that apply across all of agriculture, including vertical farming.
As with many Whats the Difference articles, the comparisons are not necessarily one-to-one, but hopefully you picked up something about ML or vertical farms that was of interest. Many issues dont map well, like problems of pollination for vertical farms. Though the output of vertical farms will likely feed some ML developers, ML is likely to play a more important part in vertical farming given the level of automation possible with sensors, robots, and ML monitoring now available.
The rest is here:
What's the Difference Between Vertical Farming and Machine Learning? - Electronic Design
All You Need to Know About Support Vector Machines – Spiceworks News and Insights
A support vector machine (SVM) is defined as a machine learning algorithm that uses supervised learning models to solve complex classification, regression, and outlier detection problems by performing optimal data transformations that determine boundaries between data points based on predefined classes, labels, or outputs. This article explains the fundamentals of SVMs, their working, types, and a few real-world examples.
A support vector machine (SVM) is a machine learning algorithm that uses supervised learning models to solve complex classification, regression, and outlier detection problems by performing optimal data transformations that determine boundaries between data points based on predefined classes, labels, or outputs. SVMs are widely adopted across disciplines such as healthcare, natural language processing, signal processing applications, and speech & image recognition fields.
Technically, the primary objective of the SVM algorithm is to identify a hyperplane that distinguishably segregates the data points of different classes. The hyperplane is localized in such a manner that the largest margin separates the classes under consideration.
The support vector representation is shown in the figure below:
As seen in the above figure, the margin refers to the maximum width of the slice that runs parallel to the hyperplane without any internal support vectors. Such hyperplanes are easier to define for linearly separable problems; however, for real-life problems or scenarios, the SVM algorithm tries to maximize the margin between the support vectors, thereby giving rise to incorrect classifications for smaller sections of data points.
SVMs are potentially designed for binary classification problems. However, with the rise in computationally intensive multiclass problems, several binary classifiers are constructed and combined to formulate SVMs that can implement such multiclass classifications through binary means.
In the mathematical context, an SVM refers to a set of ML algorithms that use kernel methods to transform data features by employing kernel functions. Kernel functions rely on the process of mapping complex datasets to higher dimensions in a manner that makes data point separation easier. The function simplifies the data boundaries for non-linear problems by adding higher dimensions to map complex data points.
While introducing additional dimensions, the data is not entirely transformed as it can act as a computationally taxing process. This technique is usually referred to as the kernel trick, wherein data transformation into higher dimensions is achieved efficiently and inexpensively.
The idea behind the SVM algorithm was first captured in 1963 by Vladimir N. Vapnik and Alexey Ya. Chervonenkis. Since then, SVMs have gained enough popularity as they have continued to have wide-scale implications across several areas, including the protein sorting process, text categorization, facial recognition, autonomous cars, robotic systems, and so on.
See More: What Is a Neural Network? Definition, Working, Types, and Applications in 2022
The working of a support vector machine can be better understood through an example. Lets assume we have red and black labels with the features denoted by x and y. We intend to have a classifier for these tags that classifies data into either the red or black category.
Lets plot the labeled data on an x-y plane, as below:
A typical SVM separates these data points into red and black tags using the hyperplane, which is a two-dimensional line in this case. The hyperplane denotes the decision boundary line, wherein data points fall under the red or black category.
A hyperplane is defined as a line that tends to widen the margins between the two closest tags or labels (red and black). The distance of the hyperplane to the most immediate label is the largest, making the data classification easier.
The above scenario is applicable for linearly separable data. However, for non-linear data, a simple straight line cannot separate the distinct data points.
Heres an example of non-linear complex dataset data:
The above dataset reveals that a single hyperplane is not sufficient to separate the involved labels or tags. However, here, the vectors are visibly distinct, making segregating them easier.
For data classification, you need to add another dimension to the feature space. For linear data discussed until this point, two dimensions of x and y were sufficient. In this case, we add a z-dimension to better classify the data points. Moreover, for convenience, lets use the equation for a circle, z = x + y.
With the third dimension, the slice of feature space along the z-direction looks like this:
Now, with three dimensions, the hyperplane, in this case, runs parallel to the x-direction at a particular value of z; lets consider it as z=1.
The remaining data points are further mapped back to two dimensions.
The above figure reveals the boundary for data points along features x, y, and z along a circle of the circumference with radii of 1 unit that segregates two labels of tags via the SVM.
Lets consider another method of visualizing data points in three dimensions for separating two tags (two different colored tennis balls in this case). Consider the balls lying on a 2D plane surface. Now, if we lift the surface upward, all the tennis balls are distributed in the air. The two differently colored balls may separate in the air at one point in this process. While this occurs, you can use or place the surface between two segregated sets of balls.
In this entire process, the act of lifting the 2D surface refers to the event of mapping data into higher dimensions, which is technically referred to as kernelling, as mentioned earlier. In this way, complex data points can be separated with the help of more dimensions. The concept highlighted here is that the data points continue to get mapped into higher dimensions until a hyperplane is identified that shows a clear separation between the data points.
The figure below gives the 3D visualization of the above use case:
See More: Narrow AI vs. General AI vs. Super AI: Key Comparisons
Support vector machines are broadly classified into two types: simple or linear SVM and kernel or non-linear SVM.
A linear SVM refers to the SVM type used for classifying linearly separable data. This implies that when a dataset can be segregated into categories or classes with the help of a single straight line, it is termed a linear SVM, and the data is referred to as linearly distinct or separable. Moreover, the classifier that classifies such data is termed a linear SVM classifier.
A simple SVM is typically used to address classification and regression analysis problems.
Non-linear data that cannot be segregated into distinct categories with the help of a straight line is classified using a kernel or non-linear SVM. Here, the classifier is referred to as a non-linear classifier. The classification can be performed with a non-linear data type by adding features into higher dimensions rather than relying on 2D space. Here, the newly added features fit a hyperplane that helps easily separate classes or categories.
Kernel SVMs are typically used to handle optimization problems that have multiple variables.
See More: What is Sentiment Analysis? Definition, Tools, and Applications
SVMs rely on supervised learning methods to classify unknown data into known categories. These find applications in diverse fields.
Here, well look at some of the top real-world examples of SVMs:
The geo-sounding problem is one of the widespread use cases for SVMs, wherein the process is employed to track the planets layered structure. This entails solving the inversion problems where the observations or results of the issues are used to factor in the variables or parameters that produced them.
In the process, linear function and support vector algorithmic models separate the electromagnetic data. Moreover, linear programming practices are employed while developing the supervised models in this case. As the problem size is considerably small, the dimension size is inevitably tiny, which accounts for mapping the planets structure.
Soil liquefaction is a significant concern when events such as earthquakes occur. Assessing its potential is crucial while designing any civil infrastructure. SVMs play a key role in determining the occurrence and non-occurrence of such liquefaction aspects. Technically, SVMs handle two tests: SPT (Standard Penetration Test) and CPT (Cone Penetration Test), which use field data to adjudicate the seismic status.
Moreover, SVMs are used to develop models that involve multiple variables, such as soil factors and liquefaction parameters, to determine the ground surface strength. It is believed that SVMs achieve an accuracy of close to 96-97% for such applications.
Protein remote homology is a field of computational biology where proteins are categorized into structural and functional parameters depending on the sequence of amino acids when sequence identification is seemingly difficult. SVMs play a key role in remote homology, with kernel functions determining the commonalities between protein sequences.
Thus, SVMs play a defining role in computational biology.
SVMs are known to solve complex mathematical problems. However, smooth SVMs are preferred for data classification purposes, wherein smoothing techniques that reduce the data outliers and make the pattern identifiable are used.
Thus, for optimization problems, smooth SVMs use algorithms such as the Newton-Armijo algorithm to handle larger datasets that conventional SVMs cannot. Smooth SVM types typically explore math properties such as strong convexity for more straightforward data classification, even with non-linear data.
SVMs classify facial structures vs. non-facial ones. The training data uses two classes of face entity (denoted by +1) and non-face entity (denoted as -1) and n*n pixels to distinguish between face and non-face structures. Further, each pixel is analyzed, and the features from each one are extracted that denote face and non-face characters. Finally, the process creates a square decision boundary around facial structures based on pixel intensity and classifies the resultant images.
Moreover, SVMs are also used for facial expression classification, which includes expressions denoted as happy, sad, angry, surprised, and so on.
In the current scenario, SVMs are used for the classification of images of surfaces. Implying that the images clicked of surfaces can be fed into SVMs to determine the texture of surfaces in those images and classify them as smooth or gritty surfaces.
Text categorization refers to classifying data into predefined categories. For example, news articles contain politics, business, the stock market, or sports. Similarly, one can segregate emails into spam, non-spam, junk, and others.
Technically, each article or document is assigned a score, which is then compared to a predefined threshold value. The article is classified into its respective category depending on the evaluated score.
For handwriting recognition examples, the dataset containing passages that different individuals write is supplied to SVMs. Typically, SVM classifiers are trained with sample data initially and are later used to classify handwriting based on score values. Subsequently, SVMs are also used to segregate writings by humans and computers.
In speech recognition examples, words from speeches are individually picked and separated. Further, for each word, certain features and characteristics are extracted. Feature extraction techniques include Mel Frequency Cepstral Coefficients (MFCC), Linear Prediction Coefficients (LPC), Linear Prediction Cepstral Coefficients (LPCC), and others.
These methods collect audio data, feed it to SVMs and then train the models for speech recognition.
With SVMs, you can determine whether any digital image is tampered with, contaminated, or pure. Such examples are helpful when handling security-related matters for organizations or government agencies, as it is easier to encrypt and embed data as a watermark in high-resolution images.
Such images contain more pixels; hence, it can be challenging to spot hidden or watermarked messages. However, one solution is to separate each pixel and store data in different datasets that SVMs can later analyze.
Medical professionals, researchers, and scientists worldwide have been toiling hard to find a solution that can effectively detect cancer in its early stages. Today, several AI and ML tools are being deployed for the same. For example, in January 2020, Google developed an AI tool that helps in early breast cancer detection and reduces false positives and negatives.
In such examples, SVMs can be employed, wherein cancerous images can be supplied as input. SVM algorithms can analyze them, train the models, and eventually categorize the images that reveal malign or benign cancer features.
See More: What Is a Decision Tree? Algorithms, Template, Examples, and Best Practices
SVMs are crucial while developing applications that involve the implementation of predictive models. SVMs are easy to comprehend and deploy. They offer a sophisticated machine learning algorithm to process linear and non-linear data through kernels.
SVMs find applications in every domain and real-life scenarios where data is handled by adding higher dimensional spaces. This entails considering factors such as the tuning hyper-parameters, selecting the kernel for execution, and investing time and resources in the training phase, which help develop the supervised learning models.
Did this article help you understand the concept of support vector machines? Comment below or let us know on Facebook, Twitter, or LinkedIn. Wed love to hear from you!
Here is the original post:
All You Need to Know About Support Vector Machines - Spiceworks News and Insights
Solve the problem of unstructured data with machine learning – VentureBeat
Were you unable to attend Transform 2022? Check out all of the summit sessions in our on-demand library now! Watch here.
Were in the midst of a data revolution. The volume of digital data created within the next five years will total twice the amount produced so far and unstructured data will define this new era of digital experiences.
Unstructured data information that doesnt follow conventional models or fit into structured database formats represents more than 80% of all new enterprise data. To prepare for this shift, companies are finding innovative ways to manage, analyze and maximize the use of data in everything from business analytics to artificial intelligence (AI). But decision-makers are also running into an age-old problem: How do you maintain and improve the quality of massive, unwieldy datasets?
With machine learning (ML), thats how. Advancements in ML technology now enable organizations to efficiently process unstructured data and improve quality assurance efforts. With a data revolution happening all around us, where does your company fall? Are you saddled with valuable, yet unmanageable datasets or are you using data to propel your business into the future?
Theres no disputing the value of accurate, timely and consistent data for modern enterprises its as vital as cloud computing and digital apps. Despite this reality, however, poor data quality still costs companies an average of $13 million annually.
MetaBeat 2022
MetaBeat will bring together thought leaders to give guidance on how metaverse technology will transform the way all industries communicate and do business on October 4 in San Francisco, CA.
To navigate data issues, you may apply statistical methods to measure data shapes, which enables your data teams to track variability, weed out outliers, and reel in data drift. Statistics-based controls remain valuable to judge data quality and determine how and when you should turn to datasets before making critical decisions. While effective, this statistical approach is typically reserved for structured datasets, which lend themselves to objective, quantitative measurements.
But what about data that doesnt fit neatly into Microsoft Excel or Google Sheets, including:
When these types of unstructured data are at play, its easy for incomplete or inaccurate information to slip into models. When errors go unnoticed, data issues accumulate and wreak havoc on everything from quarterly reports to forecasting projections. A simple copy and paste approach from structured data to unstructured data isnt enough and can actually make matters much worse for your business.
The common adage, garbage in, garbage out, is highly applicable in unstructured datasets. Maybe its time to trash your current data approach.
When considering solutions for unstructured data, ML should be at the top of your list. Thats because ML can analyze massive datasets and quickly find patterns among the clutter and with the right training, ML models can learn to interpret, organize and classify unstructured data types in any number of forms.
For example, an ML model can learn to recommend rules for data profiling, cleansing and standardization making efforts more efficient and precise in industries like healthcare and insurance. Likewise, ML programs can identify and classify text data by topic or sentiment in unstructured feeds, such as those on social media or within email records.
As you improve your data quality efforts through ML, keep in mind a few key dos and donts:
Your unstructured data is a treasure trove for new opportunities and insights. Yet only 18% of organizations currently take advantage of their unstructured data and data quality is one of the top factors holding more businesses back.
As unstructured data becomes more prevalent and more pertinent to everyday business decisions and operations, ML-based quality controls provide much-needed assurance that your data is relevant, accurate, and useful. And when you arent hung up on data quality, you can focus on using data to drive your business forward.
Just think about the possibilities that arise when you get your data under control or better yet, let ML take care of the work for you.
Edgar Honing is senior solutions architect at AHEAD.
Welcome to the VentureBeat community!
DataDecisionMakers is where experts, including the technical people doing data work, can share data-related insights and innovation.
If you want to read about cutting-edge ideas and up-to-date information, best practices, and the future of data and data tech, join us at DataDecisionMakers.
You might even considercontributing an articleof your own!
Read More From DataDecisionMakers
More here:
Solve the problem of unstructured data with machine learning - VentureBeat
Ray, the machine learning tech behind OpenAI, levels up to Ray 2.0 – VentureBeat
Were you unable to attend Transform 2022? Check out all of the summit sessions in our on-demand library now! Watch here.
Over the last two years, one of the most common ways for organizations to scale and run increasingly large and complex artificial intelligence (AI) workloads has been with the open-source Ray framework, used by companies from OpenAI to Shopify and Instacart.
Ray enables machine learning (ML) models to scale across hardware resources and can also be used to support MLops workflows across different ML tools. Ray 1.0 came out in September 2020 and has had a series of iterations over the last two years.
Today, the next major milestone was released, with the general availability of Ray 2.0 at the Ray Summit in San Francisco. Ray 2.0 extends the technology with the new Ray AI Runtime (AIR) that is intended to work as a runtime layer for executing ML services. Ray 2.0 also includes capabilities designed to help simplify building and managing AI workloads.
Alongside the new release, Anyscale, which is the lead commercial backer of Ray, announced a new enterprise platform for running Ray. Anyscale also announced a new $99 million round of funding co-led by existing investors Addition and Intel Capital with participation from Foundation Capital.
MetaBeat 2022
MetaBeat will bring together thought leaders to give guidance on how metaverse technology will transform the way all industries communicate and do business on October 4 in San Francisco, CA.
Ray started as a small project at UC Berkeley and it has grown far beyond what we imagined at the outset, said Robert Nishihara, cofounder and CEO at Anyscale, during his keynote at the Ray Summit.
Its hard to understate the foundational importance and reach of Ray in the AI space today.
Nishihara went through a laundry list of big names in the IT industry that are using Ray during his keynote. Among the companies he mentioned is ecommerce platform vendor Shopify, which uses Ray to help scale its ML platform that makes use of TensorFlow and PyTorch. Grocery delivery service Instacart is another Ray user, benefitting from the technology to help train thousands of ML models. Nishihara noted that Amazon is also a Ray user across multiple types of workloads.
Ray is also a foundational element for OpenAI, which is one of the leading AI innovators, and is the group behind the GPT-3 Large Language Model and DALL-E image generation technology.
Were using Ray to train our largest models, Greg Brockman, CTO and cofounder of OpenAI, said at the Ray Summit. So, it has been very helpful for us in terms of just being able to scale up to a pretty unprecedented scale.
Brockman commented that he sees Ray as a developer-friendly tool and the fact that it is a third-party tool that OpenAI doesnt have to maintain is helpful, too.
When something goes wrong, we can complain on GitHub and get an engineer to go work on it, so it reduces some of the burden of building and maintaining infrastructure, Brockman said.
For Ray 2.0, a primary goal for Nishihara was to make it simpler for more users to be able to benefit from the technology, while providing performance optimizations that benefit users big and small.
Nishihara commented that a common pain point in AI is that organizations can get tied into a particular framework for a certain workload, but realize over time they also want to use other frameworks. For example, an organization might start out just using TensorFlow, but realize they also want to use PyTorch and HuggingFace in the same ML workload. With the Ray AI Runtime (AIR) in Ray 2.0, it will now be easier for users to unify ML workloads across multiple tools.
Model deployment is another common pain point that Ray 2.0 is looking to help solve, with the Ray Serve deployment graph capability.
Its one thing to deploy a handful of machine learning models. Its another thing entirely to deploy several hundred machine learning models, especially when those models may depend on each other and have different dependencies, Nishihara said. As part of Ray 2.0, were announcing Ray Serve deployment graphs, which solve this problem and provide a simple Python interface for scalable model composition.
Looking forward, Nishiharas goal with Ray is to help enable a broader use of AI by making it easier to develop and manage ML workloads.
Wed like to get to the point where any developer or any organization can succeed with AI and get value from AI, Nishihara said.
VentureBeat's mission is to be a digital town square for technical decision-makers to gain knowledge about transformative enterprise technology and transact. Learn more about membership.
Read the original here:
Ray, the machine learning tech behind OpenAI, levels up to Ray 2.0 - VentureBeat
Wind – Machine learning and AI specialist Cognitive Business collaborates with Weatherquest on weather forecasts for offshore wind platform -…
A data driven tool that predicts with 99.9 percent accuracy the safest and most successful windows for crew transfers to offshore wind platforms WAVES is the first technology of its kind and is already being used by RWE across its Robin Rigg and Rampion windfarms.
The collaboration has now seen RWE integrate Weatherquests API into the already operational WAVES platform on Robin Rigg to work alongside other forecast data to enable in-day and week-ahead O&M decision-support for turbine specific accessibility.
The integration of WAVES with Weatherquests API allows us to develop our unique technology yet further to make it an even more trusted tool for windfarm owners and operators to plan and schedule their O&M programmes said MD at Cognitive Business, Ty Burridge Oakland, speaking about the upgrades to its WAVES technology. WAVES is already a hugely accurate and relied upon technology in the industry for effectively, efficiently and safely deploying crews onto windfarms to conduct repairs and maintenance and by integrating weather forecast data, we can confidently say we have made an already highly valued technology an even more robust tool for managing and planning offshore wind repair and maintenance programmes.
Developed by Nottingham and London based, Cognitive Business in 2020, WAVES was funded in the same year by the Offshore Wind Growth Partnership to better predict safer and more successful windows for crew transfers to offshore wind platforms.
Steve Dorling, Chief Executive at Weatherquest, added that WAVES has developed a reputation within the offshore wind industry, over a number of years, for enabling owners and operators to deploy their crews with real accuracy and has been working to great effect on some of the UKs largest windfarms.
It therefore made absolute sense for us both, as data analysis businesses focused on supporting safety and productivity, to combine our expertise in this innovative way said Mr Dorling. Its great that we can further enhance the WAVES technology together in a market where it is already a trusted technology for identifying optimal windows for offshore wind crew transfers.
Cognitive Business is an industry leader in machine learning and applied A.I, developing a wide range of decision support, performance monitoring, and predictive maintenance solutions for offshore wind operations and maintenance applications.
Weatherquest is a privately owned weather forecasting and weather analysis company headquartered at the University of East Anglia providing weather forecasting support services across sectors in the UK and Northern Europe including onshore and offshore wind energy and ports.
For additional information:
Cognitive Business
Weatherquest
‘Machine Learning’ Predicts The Future With More Reliable Diagnostics – Nation World News
Headquarters of the Council of Higher Scientific Research (CSIC).
a bone scan Every two years for all women aged 50-69. Since 1990, that is The biggest testing challenge for the national health systemAnd it aims to prevent one of the most common cancers in Spain, that is Mother, The method is X-rays that detect potentially cancerous areas; If something suspicious is found, that test is followed by more tests, often High probability of false positives, harmful and costly,
they are curvature This is the main reason why screening is limited to the highest risk groups. By adding predictive algorithms to mammograms, the risk areas of a patients breasts would be limited and the reliability of diagnosis increased to 90 percent. Therefore, they can be done with Often and the age range of the women they target Expansion,
It is a process that already exists, which uses artificial intelligenceand that . develops a team of Superior Council of Scientific Inquiry (CSIC), specifically the Institute of Corpuscular Physics (IFIC). it is part of the scope of machine learning (machine learning) in precision medicine, and a research network that seeks to increase the efficiency with which each patient is treated and optimize health care resources.
To understand how, you must first understand the concepts that come into play. The first is artificial intelligence. the ability of a computer or robot to perform tasks normally associated with intelligent beings, defined as sara degli-apostic You carlos sierra, author of the CSIC white paper on the subject. That is, they are the processes that are used replace human work with robotsWith the aim of accomplishing this with greater accuracy and greater efficiency.
And where can artificial intelligence work in medicine today? On many fronts, he replies. dolores del castilloResearchers from CSICs Center for Automation and Robotics, From the administrative to the management of clinical documentation. And, in a more specific way, in the analysis of images, or in the monitoring and follow-up of patients. And where are the still bigger limits? Above all, in the field of health care, in legal and ethical aspects when dealing with important matters. And whats more, theres still a long way to go, explains Del Castillo, who works on the projects, among others. neurological movement disorderTraining for a large section of healthcare workers.
We find the second concept as a subfield of artificial intelligence, along with its advantages and disadvantages: machine learning, This can be translated as machine learning. That is, artificial intelligence that works through computers thatand find patterns in population groups, With these patterns, predictions are made about what is most likely to happen. machine learning translate data Algorithm,
Precision medicine to predict disease
and after artificial intelligence and machine learningThere is a third concept: the precision medicine, The one that suits the person, his genes, his background, his lifestyle, his socialization. a model that must first be able predictable disease, Second, Francisco Albiol from IFIC, continues to assess each patient, apply the best treatment based on clinical evidence, identify the most complex cases, and assess their inclusion in management programs.
It makes sense high impact disease, and does not make sense for serious diseases; For example, distinguishing the flu from a cold in primary care, as the benefits will not compensate for the effort required.
The key to the use of artificial intelligence in medicine is also cost optimization, which is very important for public health. Spains population has increased from 42 to 47 million people between 2003 and 2022, that is, more than 10 percent. and from 2005 to 2022, The average age of the population has increased from 40 to 44, We are getting older and older.
Therefore, Dolores del Castillo says, the best valued projects and, therefore, likely to be funded, are those that incorporate artificial intelligence techniques to address the prevention, diagnosis and treatment of cardiovascular diseases, neurodegenerative diseases, cancer and obesity. There is also a special focus on personal and home medicine, elderly care, and new drug offerings. The need for healthcare has been heightened by our demographics, and The aim should be to reduce and simplify the challenges with technology, we tried machine learning, summarizes Albiol.
Albiol is one of the scientists who led a program to improve breast cancer detection through algorithms. He defends, like other researchers, that if we mix machine learning with precision medicine, we should be talking about 4p medicine. Which brings together four features: Predictive, personal, preventive and participatory,
Because most purists confine precision medicine to the field of patient genetics, and would not include it in the bag that takes more characteristics into account. Those who do say that we are talking about something much broader: Applied to precision medicine, machine learning allows for Analyze large amounts of very different types of data (genomic, biochemical, social, medical imaging) and model them to be able to offer together individual diagnosismore precise and thus more effective treatment, summarizes researcher Lara Loret Iglesias of the Institute of Physics of Cantabria.
Lloret is part of a network of scientists who, like Albiol or Del Castillo, are dedicated to projects on machine learning and precision medicine. One of them developed by his team, which he leads together with fellow physicist Miriam Kobo Cano, is called Branyas. It is in honor of Spains oldest woman, Maria Branyas, who managed to overcome Covid-19: she has done so at the age of 113. In this they bring together the many casuistries of more than 3,000 elderly people, much less just genetics: machine learning establish Risk profile of getting sick or dying as a result of coronavirus, We derived data from the analysis of three risk profiles: a sociodemographic, a biological and an extended biological, which will add information on issues such as aspects related to the intestinal microbiota, vaccination and immunity.
Precision Medicine, Cancer and Alzheimers
also explain this Joseph Lewis Arcosfrom the Artificial Intelligence Research Institute. common diseases There are cancer and Alzheimers linked to precision medicine, but they have stood out with the Ictus project. Launched in the middle of a pandemic (which has made things difficult, he admits), he has treated patients at Barcelonas Belwitz Hospital who suffered strokes and, after a severe and acute phase, Have become long term,
In particular, those with movement difficulty in one hand or both. made over 700 sessions In which patients have been asked to play the keyboard of the electronic piano. Then, they transferred the analysis of finger movements to the results to see what the patterns of difficulties and improvements are. And theyve gotten particularly positive feedback among users because its not only doing an exercise, but it affects a very emotional part. The goal is now to expand it to hospitals in the United Kingdom.,
and future? Dolores del Castillo replies, I believe that the challenge of artificial intelligence in medicine is to incorporate research results into daily practice in a generalized way, but always without forgetting that it is the experts who have is the last word. To do that, doctors need to be able to rely on these systems and Interact with them in the most natural and simple wayEven helping with its design.
Lara Loret believes that we have to be able to build generalizable prediction systems, that is, the efficiency of the model does not depend on unnecessary things such as which machine the data is taken in, or how the calibration is. Francisco Albiol focuses on a problem that may be in the long run must have a solutionAt present, larger hospitals are preferred in these technologies than smaller cities or towns. convenience and reduce costs It also has to do with reaching out to everyone.
While it may include statements, data or notes from health institutions or professionals, the information contained in medical writing is edited and prepared by journalists. We advise the reader to consult a health professional on any health-related questions.
Continue reading here:
'Machine Learning' Predicts The Future With More Reliable Diagnostics - Nation World News
Uber and AMC bring machine learning company Rokt onboard to drive revenue – Mugglehead
Rokt has partnered with Uber Technologies (NYSE:UBER) and AMC Theatres (NYSE:AMC) to help both companies make more money on their websites and mobile apps.
Rokt is an ecommerce tech company using machine learning to help tailor transactions to each shopper. The idea behind the technology is to give companies the chance to get additional revenue, find customers at scale and give extra options to existing customers by using machine learning to present offers to each shopper as theyre entering into the final stages of a transaction. The analog here would be the impulse buying section prior to a checkout line, except specifically tailored due to each consumer due to collected data.
Uber and AMC Theatres are two of the most recognized brands in the world and were extremely pleased to partner with both of them as we accelerate our growth globally. Our global partnership with Uber will support the Uber Eats internal ad network and unlock additional profitability for the company. Our partnership with AMC has already begun generating outstanding results for the company. We look forward to expanding our relationships with both of these companies in the future, said Elizabeth Buchanan, chief commercial officer of Rokt.
Rokts deal is ecommerce technology that helps customers find the full potential of every transaction to grow revenue. Existing customers include Live Nation, Groupon, Staples, Lands End, Fanatics, GoDaddy, Vistaprint and HelloFresh, but also extend out to include 2,500 other global businesses and advertisers. The company is originally from Australia, but its moved its headquarters to New York City in the United States, and has expanded out to include 19 countries across three continents.
Rokts partnership with Uber will initially launch with Uber Eats in the US, Canada, Australia and Japan, with Rokts machine learning technology driving additional revenue for Uber during the checkout experience. AMC has partnered with Rokt to drive revenue and customer lifetime value across the companys online and mobile channels.
As millions of moviegoers come to AMC each week to enjoy the unmatched entertainment of the big screen, its important that we are offering a guest experience thats personally relevant across the entire moviegoing journey. Our partnership with Rokt enables us to better personally engage our consumers and drive higher value per transaction by optimizing each online touchpoint without adding additional cost to the moviegoer, said Mark Pearson, chief strategy officer for AMC Theatres.
Rokt uses intelligence taken from five billion transactions across hundreds of ecommerce businesses to allow brands to create a tailored customer experience wherein they can control the types of offers on display to their customers. Businesses that partner with Rokt can unlock profit upwards to $0.30 per transaction through high performance techniques relevant to each individual from the moment where the customer puts the item in their digital cart to the time their payment goes through.
Read the original here:
Uber and AMC bring machine learning company Rokt onboard to drive revenue - Mugglehead