---
_id: '999'
abstract:
- lang: eng
text: 'In multi-task learning, a learner is given a collection of prediction tasks
and needs to solve all of them. In contrast to previous work, which required that
annotated training data must be available for all tasks, we consider a new setting,
in which for some tasks, potentially most of them, only unlabeled training data
is provided. Consequently, to solve all tasks, information must be transferred
between tasks with labels and tasks without labels. Focusing on an instance-based
transfer method we analyze two variants of this setting: when the set of labeled
tasks is fixed, and when it can be actively selected by the learner. We state
and prove a generalization bound that covers both scenarios and derive from it
an algorithm for making the choice of labeled tasks (in the active case) and for
transferring information between the tasks in a principled way. We also illustrate
the effectiveness of the algorithm on synthetic and real data. '
alternative_title:
- PMLR
article_processing_charge: No
author:
- first_name: Anastasia
full_name: Pentina, Anastasia
id: 42E87FC6-F248-11E8-B48F-1D18A9856A87
last_name: Pentina
- first_name: Christoph
full_name: Lampert, Christoph
id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
last_name: Lampert
orcid: 0000-0001-8622-7887
citation:
ama: 'Pentina A, Lampert C. Multi-task learning with labeled and unlabeled tasks.
In: Vol 70. ML Research Press; 2017:2807-2816.'
apa: 'Pentina, A., & Lampert, C. (2017). Multi-task learning with labeled and
unlabeled tasks (Vol. 70, pp. 2807–2816). Presented at the ICML: International
Conference on Machine Learning, Sydney, Australia: ML Research Press.'
chicago: Pentina, Anastasia, and Christoph Lampert. “Multi-Task Learning with Labeled
and Unlabeled Tasks,” 70:2807–16. ML Research Press, 2017.
ieee: 'A. Pentina and C. Lampert, “Multi-task learning with labeled and unlabeled
tasks,” presented at the ICML: International Conference on Machine Learning, Sydney,
Australia, 2017, vol. 70, pp. 2807–2816.'
ista: 'Pentina A, Lampert C. 2017. Multi-task learning with labeled and unlabeled
tasks. ICML: International Conference on Machine Learning, PMLR, vol. 70, 2807–2816.'
mla: Pentina, Anastasia, and Christoph Lampert. Multi-Task Learning with Labeled
and Unlabeled Tasks. Vol. 70, ML Research Press, 2017, pp. 2807–16.
short: A. Pentina, C. Lampert, in:, ML Research Press, 2017, pp. 2807–2816.
conference:
end_date: 2017-08-11
location: Sydney, Australia
name: 'ICML: International Conference on Machine Learning'
start_date: 2017-08-06
date_created: 2018-12-11T11:49:37Z
date_published: 2017-06-08T00:00:00Z
date_updated: 2023-10-17T11:53:32Z
day: '08'
department:
- _id: ChLa
ec_funded: 1
external_id:
isi:
- '000683309502093'
intvolume: ' 70'
isi: 1
language:
- iso: eng
main_file_link:
- open_access: '1'
url: https://arxiv.org/abs/1602.06518
month: '06'
oa: 1
oa_version: Submitted Version
page: 2807 - 2816
project:
- _id: 2532554C-B435-11E9-9278-68D0E5697425
call_identifier: FP7
grant_number: '308036'
name: Lifelong Learning of Visual Scene Understanding
publication_identifier:
isbn:
- '9781510855144'
publication_status: published
publisher: ML Research Press
publist_id: '6399'
quality_controlled: '1'
scopus_import: '1'
status: public
title: Multi-task learning with labeled and unlabeled tasks
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 70
year: '2017'
...
---
_id: '1098'
abstract:
- lang: eng
text: Better understanding of the potential benefits of information transfer and
representation learning is an important step towards the goal of building intelligent
systems that are able to persist in the world and learn over time. In this work,
we consider a setting where the learner encounters a stream of tasks but is able
to retain only limited information from each encountered task, such as a learned
predictor. In contrast to most previous works analyzing this scenario, we do not
make any distributional assumptions on the task generating process. Instead, we
formulate a complexity measure that captures the diversity of the observed tasks.
We provide a lifelong learning algorithm with error guarantees for every observed
task (rather than on average). We show sample complexity reductions in comparison
to solving every task in isolation in terms of our task complexity measure. Further,
our algorithmic framework can naturally be viewed as learning a representation
from encountered tasks with a neural network.
acknowledgement: "This work was in parts funded by the European Research Council under
the European Union’s Seventh Framework Programme (FP7/2007-2013)/ERC grant agreement
no 308036.\r\n\r\n"
alternative_title:
- Advances in Neural Information Processing Systems
author:
- first_name: Anastasia
full_name: Pentina, Anastasia
id: 42E87FC6-F248-11E8-B48F-1D18A9856A87
last_name: Pentina
- first_name: Ruth
full_name: Urner, Ruth
last_name: Urner
citation:
ama: 'Pentina A, Urner R. Lifelong learning with weighted majority votes. In: Vol
29. Neural Information Processing Systems; 2016:3619-3627.'
apa: 'Pentina, A., & Urner, R. (2016). Lifelong learning with weighted majority
votes (Vol. 29, pp. 3619–3627). Presented at the NIPS: Neural Information Processing
Systems, Barcelona, Spain: Neural Information Processing Systems.'
chicago: Pentina, Anastasia, and Ruth Urner. “Lifelong Learning with Weighted Majority
Votes,” 29:3619–27. Neural Information Processing Systems, 2016.
ieee: 'A. Pentina and R. Urner, “Lifelong learning with weighted majority votes,”
presented at the NIPS: Neural Information Processing Systems, Barcelona, Spain,
2016, vol. 29, pp. 3619–3627.'
ista: 'Pentina A, Urner R. 2016. Lifelong learning with weighted majority votes.
NIPS: Neural Information Processing Systems, Advances in Neural Information Processing
Systems, vol. 29, 3619–3627.'
mla: Pentina, Anastasia, and Ruth Urner. Lifelong Learning with Weighted Majority
Votes. Vol. 29, Neural Information Processing Systems, 2016, pp. 3619–27.
short: A. Pentina, R. Urner, in:, Neural Information Processing Systems, 2016, pp.
3619–3627.
conference:
end_date: 2016-12-10
location: Barcelona, Spain
name: 'NIPS: Neural Information Processing Systems'
start_date: 2016-12-05
date_created: 2018-12-11T11:50:08Z
date_published: 2016-12-01T00:00:00Z
date_updated: 2021-01-12T06:48:15Z
day: '01'
ddc:
- '006'
department:
- _id: ChLa
ec_funded: 1
file:
- access_level: open_access
content_type: application/pdf
creator: system
date_created: 2018-12-12T10:12:42Z
date_updated: 2018-12-12T10:12:42Z
file_id: '4961'
file_name: IST-2017-775-v1+1_main.pdf
file_size: 237111
relation: main_file
- access_level: open_access
content_type: application/pdf
creator: system
date_created: 2018-12-12T10:12:43Z
date_updated: 2018-12-12T10:12:43Z
file_id: '4962'
file_name: IST-2017-775-v1+2_supplementary.pdf
file_size: 185818
relation: main_file
file_date_updated: 2018-12-12T10:12:43Z
has_accepted_license: '1'
intvolume: ' 29'
language:
- iso: eng
month: '12'
oa: 1
oa_version: Published Version
page: 3619-3627
project:
- _id: 2532554C-B435-11E9-9278-68D0E5697425
call_identifier: FP7
grant_number: '308036'
name: Lifelong Learning of Visual Scene Understanding
publication_status: published
publisher: Neural Information Processing Systems
publist_id: '6277'
pubrep_id: '775'
quality_controlled: '1'
scopus_import: 1
status: public
title: Lifelong learning with weighted majority votes
type: conference
user_id: 3E5EF7F0-F248-11E8-B48F-1D18A9856A87
volume: 29
year: '2016'
...
---
_id: '1126'
abstract:
- lang: eng
text: "Traditionally machine learning has been focusing on the problem of solving
a single\r\ntask in isolation. While being quite well understood, this approach
disregards an\r\nimportant aspect of human learning: when facing a new problem,
humans are able to\r\nexploit knowledge acquired from previously learned tasks.
Intuitively, access to several\r\nproblems simultaneously or sequentially could
also be advantageous for a machine\r\nlearning system, especially if these tasks
are closely related. Indeed, results of many\r\nempirical studies have provided
justification for this intuition. However, theoretical\r\njustifications of this
idea are rather limited.\r\nThe focus of this thesis is to expand the understanding
of potential benefits of information\r\ntransfer between several related learning
problems. We provide theoretical\r\nanalysis for three scenarios of multi-task
learning - multiple kernel learning, sequential\r\nlearning and active task selection.
We also provide a PAC-Bayesian perspective on\r\nlifelong learning and investigate
how the task generation process influences the generalization\r\nguarantees in
this scenario. In addition, we show how some of the obtained\r\ntheoretical results
can be used to derive principled multi-task and lifelong learning\r\nalgorithms
and illustrate their performance on various synthetic and real-world datasets."
acknowledgement: "First and foremost I would like to express my gratitude to my supervisor,
Christoph\r\nLampert. Thank you for your patience in teaching me all aspects of
doing research\r\n(including English grammar), for your trust in my capabilities
and endless support. Thank\r\nyou for granting me freedom in my research and, at
the same time, having time and\r\nhelping me cope with the consequences whenever
I needed it. Thank you for creating\r\nan excellent atmosphere in the group, it
was a great pleasure and honor to be a part of\r\nit. There could not have been
a better and more inspiring adviser and mentor.\r\nI thank Shai Ben-David for welcoming
me into his group at the University of Waterloo,\r\nfor inspiring discussions and
support. It was a great pleasure to work together. I am\r\nalso thankful to Ruth
Urner for hosting me at the Max-Planck Institute Tübingen, for the\r\nfruitful
collaboration and for taking care of me during that not-so-sunny month of May.\r\nI
thank Jan Maas for kindly joining my thesis committee despite the short notice and\r\nproviding
me with insightful comments.\r\nI would like to thank my colleagues for their support,
entertaining conversations and\r\nendless table soccer games we shared together:
Georg, Jan, Amelie and Emilie, Michal\r\nand Alex, Alex K. and Alex Z., Thomas,
Sameh, Vlad, Mayu, Nathaniel, Silvester, Neel,\r\nCsaba, Vladimir, Morten. Thank
you, Mabel and Ram, for the wonderful time we spent\r\ntogether. I am thankful to
Shrinu and Samira for taking care of me during my stay at the\r\nUniversity of Waterloo.
Special thanks to Viktoriia for her never-ending optimism and for\r\nbeing so inspiring
and supportive, especially at the beginning of my PhD journey.\r\nThanks to IST
administration, in particular, Vlad and Elisabeth for shielding me from\r\nmost
of the bureaucratic paperwork.\r\n\r\nThis dissertation would not have been possible
without funding from the European\r\nResearch Council under the European Union's
Seventh Framework Programme\r\n(FP7/2007-2013)/ERC grant agreement no 308036."
alternative_title:
- ISTA Thesis
article_processing_charge: No
author:
- first_name: Anastasia
full_name: Pentina, Anastasia
id: 42E87FC6-F248-11E8-B48F-1D18A9856A87
last_name: Pentina
citation:
ama: Pentina A. Theoretical foundations of multi-task lifelong learning. 2016. doi:10.15479/AT:ISTA:TH_776
apa: Pentina, A. (2016). Theoretical foundations of multi-task lifelong learning.
Institute of Science and Technology Austria. https://doi.org/10.15479/AT:ISTA:TH_776
chicago: Pentina, Anastasia. “Theoretical Foundations of Multi-Task Lifelong Learning.”
Institute of Science and Technology Austria, 2016. https://doi.org/10.15479/AT:ISTA:TH_776.
ieee: A. Pentina, “Theoretical foundations of multi-task lifelong learning,” Institute
of Science and Technology Austria, 2016.
ista: Pentina A. 2016. Theoretical foundations of multi-task lifelong learning.
Institute of Science and Technology Austria.
mla: Pentina, Anastasia. Theoretical Foundations of Multi-Task Lifelong Learning.
Institute of Science and Technology Austria, 2016, doi:10.15479/AT:ISTA:TH_776.
short: A. Pentina, Theoretical Foundations of Multi-Task Lifelong Learning, Institute
of Science and Technology Austria, 2016.
date_created: 2018-12-11T11:50:17Z
date_published: 2016-11-01T00:00:00Z
date_updated: 2023-09-07T11:52:03Z
day: '01'
ddc:
- '006'
degree_awarded: PhD
department:
- _id: ChLa
doi: 10.15479/AT:ISTA:TH_776
ec_funded: 1
file:
- access_level: open_access
content_type: application/pdf
creator: system
date_created: 2018-12-12T10:14:07Z
date_updated: 2018-12-12T10:14:07Z
file_id: '5056'
file_name: IST-2017-776-v1+1_Pentina_Thesis_2016.pdf
file_size: 2140062
relation: main_file
file_date_updated: 2018-12-12T10:14:07Z
has_accepted_license: '1'
language:
- iso: eng
month: '11'
oa: 1
oa_version: Published Version
page: '127'
project:
- _id: 2532554C-B435-11E9-9278-68D0E5697425
call_identifier: FP7
grant_number: '308036'
name: Lifelong Learning of Visual Scene Understanding
publication_identifier:
issn:
- 2663-337X
publication_status: published
publisher: Institute of Science and Technology Austria
publist_id: '6234'
pubrep_id: '776'
status: public
supervisor:
- first_name: Christoph
full_name: Lampert, Christoph
id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
last_name: Lampert
orcid: 0000-0001-8622-7887
title: Theoretical foundations of multi-task lifelong learning
type: dissertation
user_id: c635000d-4b10-11ee-a964-aac5a93f6ac1
year: '2016'
...
---
_id: '1425'
abstract:
- lang: eng
text: 'In this work we aim at extending the theoretical foundations of lifelong
learning. Previous work analyzing this scenario is based on the assumption that
learning tasks are sampled i.i.d. from a task environment or limited to strongly
constrained data distributions. Instead, we study two scenarios when lifelong
learning is possible, even though the observed tasks do not form an i.i.d. sample:
first, when they are sampled from the same environment, but possibly with dependencies,
and second, when the task environment is allowed to change over time in a consistent
way. In the first case we prove a PAC-Bayesian theorem that can be seen as a direct
generalization of the analogous previous result for the i.i.d. case. For the second
scenario we propose to learn an inductive bias in form of a transfer procedure.
We present a generalization bound and show on a toy example how it can be used
to identify a beneficial transfer algorithm.'
alternative_title:
- Advances in Neural Information Processing Systems
author:
- first_name: Anastasia
full_name: Pentina, Anastasia
id: 42E87FC6-F248-11E8-B48F-1D18A9856A87
last_name: Pentina
- first_name: Christoph
full_name: Lampert, Christoph
id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
last_name: Lampert
orcid: 0000-0001-8622-7887
citation:
ama: 'Pentina A, Lampert C. Lifelong learning with non-i.i.d. tasks. In: Vol 2015.
Neural Information Processing Systems; 2015:1540-1548.'
apa: 'Pentina, A., & Lampert, C. (2015). Lifelong learning with non-i.i.d. tasks
(Vol. 2015, pp. 1540–1548). Presented at the NIPS: Neural Information Processing
Systems, Montreal, Canada: Neural Information Processing Systems.'
chicago: Pentina, Anastasia, and Christoph Lampert. “Lifelong Learning with Non-i.i.d.
Tasks,” 2015:1540–48. Neural Information Processing Systems, 2015.
ieee: 'A. Pentina and C. Lampert, “Lifelong learning with non-i.i.d. tasks,” presented
at the NIPS: Neural Information Processing Systems, Montreal, Canada, 2015, vol.
2015, pp. 1540–1548.'
ista: 'Pentina A, Lampert C. 2015. Lifelong learning with non-i.i.d. tasks. NIPS:
Neural Information Processing Systems, Advances in Neural Information Processing
Systems, vol. 2015, 1540–1548.'
mla: Pentina, Anastasia, and Christoph Lampert. Lifelong Learning with Non-i.i.d.
Tasks. Vol. 2015, Neural Information Processing Systems, 2015, pp. 1540–48.
short: A. Pentina, C. Lampert, in:, Neural Information Processing Systems, 2015,
pp. 1540–1548.
conference:
end_date: 2015-12-12
location: Montreal, Canada
name: 'NIPS: Neural Information Processing Systems'
start_date: 2015-12-07
date_created: 2018-12-11T11:51:57Z
date_published: 2015-01-01T00:00:00Z
date_updated: 2021-01-12T06:50:39Z
day: '01'
department:
- _id: ChLa
ec_funded: 1
intvolume: ' 2015'
language:
- iso: eng
main_file_link:
- open_access: '1'
url: http://papers.nips.cc/paper/6007-lifelong-learning-with-non-iid-tasks
month: '01'
oa: 1
oa_version: None
page: 1540 - 1548
project:
- _id: 2532554C-B435-11E9-9278-68D0E5697425
call_identifier: FP7
grant_number: '308036'
name: Lifelong Learning of Visual Scene Understanding
publication_status: published
publisher: Neural Information Processing Systems
publist_id: '5781'
quality_controlled: '1'
scopus_import: 1
status: public
title: Lifelong learning with non-i.i.d. tasks
type: conference
user_id: 3E5EF7F0-F248-11E8-B48F-1D18A9856A87
volume: 2015
year: '2015'
...
---
_id: '1706'
abstract:
- lang: eng
text: We consider a problem of learning kernels for use in SVM classification in
the multi-task and lifelong scenarios and provide generalization bounds on the
error of a large margin classifier. Our results show that, under mild conditions
on the family of kernels used for learning, solving several related tasks simultaneously
is beneficial over single task learning. In particular, as the number of observed
tasks grows, assuming that in the considered family of kernels there exists one
that yields low approximation error on all tasks, the overhead associated with
learning such a kernel vanishes and the complexity converges to that of learning
when this good kernel is given to the learner.
alternative_title:
- LNCS
author:
- first_name: Anastasia
full_name: Pentina, Anastasia
id: 42E87FC6-F248-11E8-B48F-1D18A9856A87
last_name: Pentina
- first_name: Shai
full_name: Ben David, Shai
last_name: Ben David
citation:
ama: 'Pentina A, Ben David S. Multi-task and lifelong learning of kernels. In: Vol
9355. Springer; 2015:194-208. doi:10.1007/978-3-319-24486-0_13'
apa: 'Pentina, A., & Ben David, S. (2015). Multi-task and lifelong learning
of kernels (Vol. 9355, pp. 194–208). Presented at the ALT: Algorithmic Learning
Theory, Banff, AB, Canada: Springer. https://doi.org/10.1007/978-3-319-24486-0_13'
chicago: Pentina, Anastasia, and Shai Ben David. “Multi-Task and Lifelong Learning
of Kernels,” 9355:194–208. Springer, 2015. https://doi.org/10.1007/978-3-319-24486-0_13.
ieee: 'A. Pentina and S. Ben David, “Multi-task and lifelong learning of kernels,”
presented at the ALT: Algorithmic Learning Theory, Banff, AB, Canada, 2015, vol.
9355, pp. 194–208.'
ista: 'Pentina A, Ben David S. 2015. Multi-task and lifelong learning of kernels.
ALT: Algorithmic Learning Theory, LNCS, vol. 9355, 194–208.'
mla: Pentina, Anastasia, and Shai Ben David. Multi-Task and Lifelong Learning
of Kernels. Vol. 9355, Springer, 2015, pp. 194–208, doi:10.1007/978-3-319-24486-0_13.
short: A. Pentina, S. Ben David, in:, Springer, 2015, pp. 194–208.
conference:
end_date: 2015-10-06
location: Banff, AB, Canada
name: 'ALT: Algorithmic Learning Theory'
start_date: 2015-10-04
date_created: 2018-12-11T11:53:35Z
date_published: 2015-01-01T00:00:00Z
date_updated: 2021-01-12T06:52:39Z
day: '01'
department:
- _id: ChLa
doi: 10.1007/978-3-319-24486-0_13
ec_funded: 1
intvolume: ' 9355'
language:
- iso: eng
main_file_link:
- open_access: '1'
url: http://arxiv.org/abs/1602.06531
month: '01'
oa: 1
oa_version: Preprint
page: 194 - 208
project:
- _id: 2532554C-B435-11E9-9278-68D0E5697425
call_identifier: FP7
grant_number: '308036'
name: Lifelong Learning of Visual Scene Understanding
publication_status: published
publisher: Springer
publist_id: '5430'
quality_controlled: '1'
scopus_import: 1
status: public
title: Multi-task and lifelong learning of kernels
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 9355
year: '2015'
...
---
_id: '1857'
abstract:
- lang: eng
text: 'Sharing information between multiple tasks enables algorithms to achieve
good generalization performance even from small amounts of training data. However,
in a realistic scenario of multi-task learning not all tasks are equally related
to each other, hence it could be advantageous to transfer information only between
the most related tasks. In this work we propose an approach that processes multiple
tasks in a sequence with sharing between subsequent tasks instead of solving all
tasks jointly. Subsequently, we address the question of curriculum learning of
tasks, i.e. finding the best order of tasks to be learned. Our approach is based
on a generalization bound criterion for choosing the task order that optimizes
the average expected classification performance over all tasks. Our experimental
results show that learning multiple related tasks sequentially can be more effective
than learning them jointly, the order in which tasks are being solved affects
the overall performance, and that our model is able to automatically discover
the favourable order of tasks. '
author:
- first_name: Anastasia
full_name: Pentina, Anastasia
id: 42E87FC6-F248-11E8-B48F-1D18A9856A87
last_name: Pentina
- first_name: Viktoriia
full_name: Sharmanska, Viktoriia
id: 2EA6D09E-F248-11E8-B48F-1D18A9856A87
last_name: Sharmanska
orcid: 0000-0003-0192-9308
- first_name: Christoph
full_name: Lampert, Christoph
id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
last_name: Lampert
orcid: 0000-0001-8622-7887
citation:
ama: 'Pentina A, Sharmanska V, Lampert C. Curriculum learning of multiple tasks.
In: IEEE; 2015:5492-5500. doi:10.1109/CVPR.2015.7299188'
apa: 'Pentina, A., Sharmanska, V., & Lampert, C. (2015). Curriculum learning
of multiple tasks (pp. 5492–5500). Presented at the CVPR: Computer Vision and
Pattern Recognition, Boston, MA, United States: IEEE. https://doi.org/10.1109/CVPR.2015.7299188'
chicago: Pentina, Anastasia, Viktoriia Sharmanska, and Christoph Lampert. “Curriculum
Learning of Multiple Tasks,” 5492–5500. IEEE, 2015. https://doi.org/10.1109/CVPR.2015.7299188.
ieee: 'A. Pentina, V. Sharmanska, and C. Lampert, “Curriculum learning of multiple
tasks,” presented at the CVPR: Computer Vision and Pattern Recognition, Boston,
MA, United States, 2015, pp. 5492–5500.'
ista: 'Pentina A, Sharmanska V, Lampert C. 2015. Curriculum learning of multiple
tasks. CVPR: Computer Vision and Pattern Recognition, 5492–5500.'
mla: Pentina, Anastasia, et al. Curriculum Learning of Multiple Tasks. IEEE,
2015, pp. 5492–500, doi:10.1109/CVPR.2015.7299188.
short: A. Pentina, V. Sharmanska, C. Lampert, in:, IEEE, 2015, pp. 5492–5500.
conference:
end_date: 2015-06-12
location: Boston, MA, United States
name: 'CVPR: Computer Vision and Pattern Recognition'
start_date: 2015-06-07
date_created: 2018-12-11T11:54:23Z
date_published: 2015-06-01T00:00:00Z
date_updated: 2023-02-23T10:17:31Z
day: '01'
department:
- _id: ChLa
doi: 10.1109/CVPR.2015.7299188
language:
- iso: eng
main_file_link:
- open_access: '1'
url: http://arxiv.org/abs/1412.1353
month: '06'
oa: 1
oa_version: Preprint
page: 5492 - 5500
publication_status: published
publisher: IEEE
publist_id: '5243'
quality_controlled: '1'
scopus_import: 1
status: public
title: Curriculum learning of multiple tasks
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
year: '2015'
...
---
_id: '2160'
abstract:
- lang: eng
text: Transfer learning has received a lot of attention in the machine learning
community over the last years, and several effective algorithms have been developed.
However, relatively little is known about their theoretical properties, especially
in the setting of lifelong learning, where the goal is to transfer information
to tasks for which no data have been observed so far. In this work we study lifelong
learning from a theoretical perspective. Our main result is a PAC-Bayesian generalization
bound that offers a unified view on existing paradigms for transfer learning,
such as the transfer of parameters or the transfer of low-dimensional representations.
We also use the bound to derive two principled lifelong learning algorithms, and
we show that these yield results comparable with existing methods.
article_processing_charge: No
author:
- first_name: Anastasia
full_name: Pentina, Anastasia
id: 42E87FC6-F248-11E8-B48F-1D18A9856A87
last_name: Pentina
- first_name: Christoph
full_name: Lampert, Christoph
id: 40C20FD2-F248-11E8-B48F-1D18A9856A87
last_name: Lampert
orcid: 0000-0001-8622-7887
citation:
ama: 'Pentina A, Lampert C. A PAC-Bayesian bound for Lifelong Learning. In: Vol
32. ML Research Press; 2014:991-999.'
apa: 'Pentina, A., & Lampert, C. (2014). A PAC-Bayesian bound for Lifelong Learning
(Vol. 32, pp. 991–999). Presented at the ICML: International Conference on Machine
Learning, Beijing, China: ML Research Press.'
chicago: Pentina, Anastasia, and Christoph Lampert. “A PAC-Bayesian Bound for Lifelong
Learning,” 32:991–99. ML Research Press, 2014.
ieee: 'A. Pentina and C. Lampert, “A PAC-Bayesian bound for Lifelong Learning,”
presented at the ICML: International Conference on Machine Learning, Beijing,
China, 2014, vol. 32, pp. 991–999.'
ista: 'Pentina A, Lampert C. 2014. A PAC-Bayesian bound for Lifelong Learning. ICML:
International Conference on Machine Learning vol. 32, 991–999.'
mla: Pentina, Anastasia, and Christoph Lampert. A PAC-Bayesian Bound for Lifelong
Learning. Vol. 32, ML Research Press, 2014, pp. 991–99.
short: A. Pentina, C. Lampert, in:, ML Research Press, 2014, pp. 991–999.
conference:
end_date: 2014-06-26
location: Beijing, China
name: 'ICML: International Conference on Machine Learning'
start_date: 2014-06-21
date_created: 2018-12-11T11:56:03Z
date_published: 2014-05-10T00:00:00Z
date_updated: 2023-10-17T11:54:24Z
day: '10'
department:
- _id: ChLa
intvolume: ' 32'
language:
- iso: eng
main_file_link:
- open_access: '1'
url: https://dl.acm.org/citation.cfm?id=3045003
month: '05'
oa: 1
oa_version: Submitted Version
page: 991 - 999
publication_status: published
publisher: ML Research Press
publist_id: '4844'
quality_controlled: '1'
scopus_import: '1'
status: public
title: A PAC-Bayesian bound for Lifelong Learning
type: conference
user_id: 2DF688A6-F248-11E8-B48F-1D18A9856A87
volume: 32
year: '2014'
...