From Learning to Meta-Learning: Reduced Training Overhead and Complexity for Communication Systems

Cited 0 time in webofscience Cited 18 time in scopus
  • Hit : 249
  • Download : 0
Machine learning methods adapt the parameters of a model, constrained to lie in a given model class, by using a fixed learning procedure based on data or active observations. Adaptation is done on a per-task basis, and retraining is needed when the system configuration changes. The resulting inefficiency in terms of data and training time requirements can be mitigated, if domain knowledge is available, by selecting a suitable model class and learning procedure, collectively known as inductive bias. However, it is generally difficult to encode prior knowledge into an inductive bias, particularly with black-box model classes such as neural networks. Meta-learning provides a way to automatize the selection of an inductive bias. Meta-learning leverages data or active observations from tasks that are expected to be related to future, and a priori unknown, tasks of interest. With a meta-trained inductive bias, training of a machine learning model can be potentially carried out with reduced training data and/or time complexity. This paper provides a high-level introduction to meta-learning with applications to communication systems.
Publisher
Institute of Electrical and Electronics Engineers Inc.
Issue Date
2020-03-17
Language
English
Citation

2nd 6G Wireless Summit, 6G SUMMIT 2020

DOI
10.1109/6GSUMMIT49458.2020.9083856
URI
http://hdl.handle.net/10203/275098
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0