Everyday AI

人工智慧,形影相隨

MediaTek

MediaTek is the world’s 4th largest global fabless semiconductor company. We are a market leader in developing innovative systems-on-chip (SoC) for mobile devices, home entertainment, connectivity and IoT products. Ultimately, we power more than 2 billion devices a year – that’s in 20 percent of homes and nearly 1 of every 3 mobile phones globally.

MediaTek Research, the AI research arm of MediaTek, brings the most advanced developments in AI to everyday devices. In addition to designing embedded AI systems, MediaTek Research contributes to the fundamental research of AI-based algorithmic solutions.

Research

Papers

LPI: Learned Positional Invariances for Transfer of Task Structure and Zero-shot Planning

Real-world tasks often include interactions with the environment where our actions can drastically change the available or desirable long-term outcomes. One formulation of this in the reinforcement learning setting is in terms of nonMarkovian rewards. Here the reward function, and thus the available rewards, are themselves history-dependent, and dynamically change given the agent-environment interactions. An important challenge for navigating such environments is to be able to capture the structure of this dynamic reward function, in a way that is interpretable and allows for optimal planning. This structure, in conjunction with the particular task setting at hand, then determines the optimal order in which actions should be executed, or subtasks completed. Planning methods face the challenge of combinatorial explosion if all such orderings need to be evaluated, however, learning invariances inherent in the task structure can alleviate this pressure. Here we propose a solution to this problem by allowing the planning method to recognise task segments where temporal ordering is irrelevant for predicting reward outcomes downstream. To facilitate this, our agent simultaneously learns to segment a task and predict the changing reward function resulting from its actions, while also learning about the permutation invariances in the its history that are relevant for this prediction. This dual approach can allow zero-shot or few-shot generalisation for complex, dynamic reinforcement learning tasks

LPI: Learned Positional Invariances for Transfer of Task Structure and Zero-shot Planning

Real-world tasks often include interactions with the environment where our actions can drastically change the available or desirable long-term outcomes. One formulation of this in the reinforcement learning setting is in terms of nonMarkovian rewards. Here the reward function, and thus the available rewards, are themselves history-dependent, and dynamically change given the agent-environment interactions. An important challenge for navigating such environments is to be able to capture the structure of this dynamic reward function, in a way that is interpretable and allows for optimal planning. This structure, in conjunction with the particular task setting at hand, then determines the optimal order in which actions should be executed, or subtasks completed. Planning methods face the challenge of combinatorial explosion if all such orderings need to be evaluated, however, learning invariances inherent in the task structure can alleviate this pressure. Here we propose a solution to this problem by allowing the planning method to recognise task segments where temporal ordering is irrelevant for predicting reward outcomes downstream. To facilitate this, our agent simultaneously learns to segment a task and predict the changing reward function resulting from its actions, while also learning about the permutation invariances in the its history that are relevant for this prediction. This dual approach can allow zero-shot or few-shot generalisation for complex, dynamic reinforcement learning tasks.

Talks

Blog

MediaTek Announces Breakthrough in Artificial Intelligence and Chip Design

This July, MediaTek will be presenting joint research with College of Electrical Engineering & Computer Science, National Taiwan University (NTU EECS) and Maxeda Technology at the ACM/IEEE Design Automation Conference (DAC), the most influential and longest-running conference on Electronic Design Automation (EDA) in the world. As an added distinction, this year’s DAC has selected the said research as one of its inaugural Publicity Papers. MediaTek collaborates with NTU EECS and Maxeda Technology to introduce a new AI-powered chip design algorithm at the ACM/IEEE Design Automation Conference The multiple-objective reinforcement learning algorithm proposed in the research can simultaneously optimize conflicting objectives in chip design in terms of power consumption, efficiency, chip area, yield, etc. It opens the door to significantly reduce development costs, cut down development time, and improve chip performance. The algorithm has already been applied to the creation of the MediaTek Dimensity mobile chip series and will be used in other product lines as well. “MediaTek strives to stay ahead of the curve in technology,” said SR Tsai, Corporate Senior Vice President of MediaTek’s Central Design Group. “To develop such cutting-edge technology, we employ artificial intelligence in certain aspects of the chip design process to make up for a…

Non-reversible Gaussian Processes

In Machine Learning, it is often convenient to use functions that are not fixed but only known with some probability. For example, let say we know the weather in Cambridge for the last few years and we would like to predict what will be the weather tomorrow (e.g. the temperature or wind speed, a function of time $f(t)$). Or, we know today’s weather in Cambridge and London, and we would like to know what is happening along the M11 motorway between the two cities. Instead of giving one predicted value, we would like a probability distribution for each time (or space point). Since time is continuous, we are looking for a distribution over functions $f(t)$. Gaussian Processes (GP) represent a simple and powerful example of distribution over functions. They are widely used for modeling and prediction (Rasmussen and Williams 2006), a few examples of GPs are shown in Figure 1. Perhaps the most important feature of a GP is the correlation function $k(t,t’)$, which determines the information we get about $f(t’)$ when measuring the true value of $f(t)$. The shape of this correlation function is the most important choice to make when we use a GP, it determines what kind…

TreeMAML

Many NLP models rely on training in one high-resource language, and they cannot be directly used to make predictions  for other languages at inference time. Most of the languages of the world are under-resourced and rely on Machine Translation (MT) to English to make use of Language Models. However, having an MT system in every direction is costly and  not the best solution for every NLP task. We propose the use of meta-learning to solve this issue. Our algorithm, TreeMAML, extends a meta-learning model, MAML [1], by exploiting hierarchical languages relationships. MAML adapts the model to each task with a few gradient steps. In our method, TreeMAML, this adaptation follows the hierarchical tree structure:In each step down the tree, gradients are pooled across language clusters: Algorithm 1.Algorithm 2 is a non-binary modification of the OTD clustering [2],  that generates the language tree without previous knowledge of the structure, allowing us to use implicit relationships between the languages.In our Experiments we adapt a high-resource language model, Multi-BERT [3], to a Few-Shot NLI task with these steps:We use the XNLI data set [4]. XNLI corpus is a crowd-sourced collection of pairs for the MultiNLI corpus with 10 different genres in 15 languages. The pairs are…

Non-reversible Gaussian Processes

In Machine Learning, it is often convenient to use functions that are not fixed but only known with some probability. For example, let say we…

TreeMAML

Many NLP models rely on training in one high-resource language, and they cannot be directly used to make predictions  for other languages at inference time. Most…

About Us

We at MediaTek Research are a dedicated research unit of the global MediaTek Group. Our mission is to explore and initiate emerging approaches in AI and ML from both fundamental and applied perspectives. We have two research centers, one in Cambridge (UK), and one in National Taiwan University. We cooperate closely with worldwide academic community.

Our team is made of researchers with a diverse background in computer science, engineering, mathematics and physics.

Contact Us

Interested in knowing more about MediaTek Research? Please feel free to contact us:

 

Our Email Address: info@mtkresearch.com