AAAI 2021: Accelerating the impact of artificial intelligence – Microsoft

The purposeoftheAssociation for the Advancement of ArtificialIntelligence, according to its bylaws, istwofold.The first is to promote researchin the area ofAI, and the second is to promote theresponsibleuseof these types of technology.Theresult was a35th AAAI Conference on Artificial Intelligence (AAAI-21)schedule thatbroadens the possibilitiesofAI andisheavily reflective ofa pivotal time in AI researchwhenexperts are asking bigger questions about how best toresponsiblydevelop, deploy, and integratethe technology.

Microsoft and its researchers have been pursuing and helping to foster responsible AI for yearsdeveloping innovative AI ethics checklists and fairness assessment tools like Fairlearn, establishing the Aether Committee to make principle-based recommendations, and laying out guidelines for human-AI interaction, to name only a few of the milestones in this area.

As a natural extension, researchers from Microsoft are presenting papers at this years AAAI that show the wide net theyre casting when it comes to developing responsible AI and using it for applications that do good. In How Linguistically Fair are Multilingual Pre-Trained Language Models?, researchers explore the fairness of current large multilingual language models across different languages. More specifically, they uncover how choices have been made about which models are fair and offer strategies for how these decision processes can be improved. Another paper demonstrates how AI can impact both specific industries and global challenges. In Where theres Smoke, theres Fire: Wildfire Risk Predictive Modeling via Historical Climate Data, researchers reexamine how AI can be used to predict wildfires by taking historical, climate, and geospatial data into account to improve modeling.

The belowselectionof AAAI-accepted papersshowcasesspecific advances with the potentialto havefar-reaching impact.AI that empowersallpeopleis the end goal, whether that bethroughbetter communication, better protection oftheirprivacy,orbetter optimization ofeverydayprocessesin specificfields.

For more on whatMicrosoft, asilver sponsor of the conference, and itsresearchers are undertaking when it comes to moving AI forward, explore more at theMicrosoft at AAAI 2021page.

In a nutshell: Reinforcement learning for order execution in quantitative investment.

Going deeper: The paper Universal Trading for Order Execution with Oracle Policy Distillation proposes a novel universal trading policy optimization framework for order execution in quantitative finance. It bridges the gap between noisy yet imperfect market states and optimal action sequences for order execution. Particularly, on one side, this framework leverages a policy distillation method that can better guide the learning of the common policy toward practically optimal execution by an oracle teacher with perfect information to approximate the optimal trading strategy. On the other side, a universal trading policy has been derived from the market data of various instruments, which is more training effective and more general to trade for different instruments.

Potential reach: This work can create an impact in the field of trading optimization in quantitative financial investment. The proposed universal learning-to-trade paradigm could substantially advance trading optimization with potentially significant profit gaining in order execution. The code is available in the Qlib project on GitHub.

First of its kind: To the best of the researchers knowledge, this is the first work to employ policy distillation in reinforcement learning to bridge the gap between imperfect noisy data and optimal action sequences. Moreover, the work shows that direct policy optimization has a great advantage over the traditional model-based financial methods and value-based model-free reinforcement learning methods.

Thepeopleand organizations involved:Kan Ren,Weiqing Liu,Dong Zhou,Jiang Bian, andTie-Yan Liufrom Microsoft ResearchAsia;Yuchen Fang,Weinan Zhang,andYong Yufrom Shanghai Jiao Tong University

Additional resources and related work:

In a nutshell: Want a translation system for languages with no written text? UWSpeech is your choice.

Going deeper: Existing speech-to-speech translation systems rely on the text of target language, and these existing systems cant be applied to unwritten target languages (languages without written text or phonemes). In the paper UWSpeech: Speech to Speech Translation for Unwritten Languages, researchers developed UWSpeech, a translation system for unwritten languages. UWSpeech converts target unwritten speech into discrete tokens with a converter. It then translates source-language speech into target discrete tokens with a translator and, finally, synthesizes target speech from target discrete tokens with an inverter. The researchers propose a method called XL-VAE in UWSpeech to enhance vector quantized variational autoencoder (VQ-VAE) with cross-lingual (XL) speech recognition, in order to train the converter and inverter of UWSpeech jointly.

Potential reach: This research sits broadly within cross-lingual speech translation, which can impact many scenarios where one spoken language needs to be translated into another. Conversations, lectures, international travel, and conferences are all examples where UWSpeech could be utilized. UWSpeech can also help to preserve unwritten languages spoken by a small amount of people.

Extended applications: Although this paper focuses on how UWSpeech can be applied to speech-to-speech translation, it can also be used to improve text-to-speech and speech-to-text translation, showing promising results in both areas. See the paper for more details.

The people and organizations involved:Xu Tan,TaoQin,andTie-Yan Liufrom the Machine Learning Group at Microsoft Research Asia;Chen Zhang, Yi Ren,andKejun Zhangfrom Zhejiang University

Additional resources and related work:

In a nutshell: Watch out! Data augmentation could actually hurt privacy. Stronger membership inference attack reveals where we need to improve protection.

Going deeper: The paper How Does Data Augmentation Affect Privacy in Machine Learning? challenges a common belief that data augmentation can prevent overfitting and hence protect the model from leakage of individual data points. The researchers developed membership inference algorithms that employ augmented instances and achieve state-of-the-art success rates of attacking well-generalized models trained with data augmentation, showing that privacy risk in these deep learning models could be greater than previously thought. Revealing this vulnerability encourages future development of techniques to strengthen the privacy protections of data augmentation as a training method.

Potential reach: The new proposed membership inference algorithms can better evaluate the privacy risk of a model and can hence help prevent other privacy attacks.

Toward better privacy: The end goal is to make a privacy guarantee in real-world machine learning tasks practical.

The people and organizations involved:Huishuai Zhang,Wei Chen,andTie-Yan Liuof Microsoft Research Asia;Da Yu,intern atMicrosoft ResearchAsiaat the time of the workand student atSunYat-Sen University;Jian Yin, Professor atSunYat-Sen University

Additional resources and related work:

In a nutshell:A two-branch convolutional neural network approach tointeractive speech and noise modeling for speech enhancement.

Going deeper: Mainstream deep learningbased speech enhancement mainly predicts speech only, ignoring the characteristics of background noises. However, traditional speech enhancement methods mostly go the opposite way, that is, they model noises with an assumption on noise distributions. The result is that their generalization capability is limited. In the paper Interactive Speech and Noise Modeling for Speech Enhancement, researchers propose the SN-Net, an interactive speech and noise modeling framework for speech enhancement, where speech and noise are simultaneously modeled in a two-branch deep neural network. Several interactions are introduced to help speech estimation benefit from noise prediction, and vice versa. As its challenging to model noises because of the diverse noise types, self-attention is employed in modeling both speech and noise. The SN-Net outperforms the state of the art by a large margin on several public datasets.

Potential reach:This technology can be widelyimpactfulin applications where speech clarity is important, including video recordings, online meetings, and virtual lessons.The research can naturally be extended to usewiththe speaker separation task (see paper for more on this).

Stateoftheart across multiple benchmarks:The researchers tested SN-Net againststate-of-the-artmodels on Voice Bank + DEMAND and the Deep Noise Suppression (DNS)challengedataset.Additionally, researchersconducted a two-speaker speech separation experiment onthe TIMITcorpus, and SN-Net outperforms Conv-TasNet, thestate-of-the-artmethod, for SDR (signal-to-distortionratio) improvement and Perceptual Evaluation of Speech Quality (PESQ). See the paper for a detailed breakdown of these tests.

The people and organizations involved:Xiulian PengandYan Lufrom the Media Computing Group at Microsoft Research Asia; Sriram Srinivasan from Microsoft; and Chengyu Zheng and Yuan Zhang from Communication University of China

Additional resourcesand related work:

Go here to see the original:
AAAI 2021: Accelerating the impact of artificial intelligence - Microsoft

Related Posts

Comments are closed.