DCOPs meet the realworld

Exploring unknown reward matrices with applications to mobile sensor networks

Manish Jain, Matthew Taylor, Milind Tambe, Makoto Yokoo

Research output: Chapter in Book/Report/Conference proceedingConference contribution

32 Citations (Scopus)

Abstract

Buoyed by recent successes in the area of distributed constraint optimization problems (DCOPs), this paper addresses challenges faced when applying DCOPs to real-world domains. Three fundamental challenges must be addressed for a class of real-world domains, requiring novel DCOP algorithms. First, agents may not know the payoff matrix and must explore the environment to determine rewards associated with variable settings. Second, agents may need to maximize total accumulated reward rather than instantaneous final reward. Third, limited time horizons disallow exhaustive exploration of the environment. We propose and implement a set of novel algorithms that combine decision-theoretic exploration approaches with DCOP-mandated coordination. In addition to simulation results, we implement these algorithms on robots, deploying DCOPs on a distributed mobile sensor network.

Original languageEnglish
Title of host publicationIJCAI-09 - Proceedings of the 21st International Joint Conference on Artificial Intelligence
Pages181-186
Number of pages6
Publication statusPublished - 2009
Event21st International Joint Conference on Artificial Intelligence, IJCAI-09 - Pasadena, CA, United States
Duration: Jul 11 2009Jul 17 2009

Other

Other21st International Joint Conference on Artificial Intelligence, IJCAI-09
CountryUnited States
CityPasadena, CA
Period7/11/097/17/09

Fingerprint

Sensor networks
Wireless networks
Robots

All Science Journal Classification (ASJC) codes

  • Artificial Intelligence

Cite this

Jain, M., Taylor, M., Tambe, M., & Yokoo, M. (2009). DCOPs meet the realworld: Exploring unknown reward matrices with applications to mobile sensor networks. In IJCAI-09 - Proceedings of the 21st International Joint Conference on Artificial Intelligence (pp. 181-186)

DCOPs meet the realworld : Exploring unknown reward matrices with applications to mobile sensor networks. / Jain, Manish; Taylor, Matthew; Tambe, Milind; Yokoo, Makoto.

IJCAI-09 - Proceedings of the 21st International Joint Conference on Artificial Intelligence. 2009. p. 181-186.

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Jain, M, Taylor, M, Tambe, M & Yokoo, M 2009, DCOPs meet the realworld: Exploring unknown reward matrices with applications to mobile sensor networks. in IJCAI-09 - Proceedings of the 21st International Joint Conference on Artificial Intelligence. pp. 181-186, 21st International Joint Conference on Artificial Intelligence, IJCAI-09, Pasadena, CA, United States, 7/11/09.
Jain M, Taylor M, Tambe M, Yokoo M. DCOPs meet the realworld: Exploring unknown reward matrices with applications to mobile sensor networks. In IJCAI-09 - Proceedings of the 21st International Joint Conference on Artificial Intelligence. 2009. p. 181-186
Jain, Manish ; Taylor, Matthew ; Tambe, Milind ; Yokoo, Makoto. / DCOPs meet the realworld : Exploring unknown reward matrices with applications to mobile sensor networks. IJCAI-09 - Proceedings of the 21st International Joint Conference on Artificial Intelligence. 2009. pp. 181-186
@inproceedings{4b13b77471154c1da6b5f3876a71d9a9,
title = "DCOPs meet the realworld: Exploring unknown reward matrices with applications to mobile sensor networks",
abstract = "Buoyed by recent successes in the area of distributed constraint optimization problems (DCOPs), this paper addresses challenges faced when applying DCOPs to real-world domains. Three fundamental challenges must be addressed for a class of real-world domains, requiring novel DCOP algorithms. First, agents may not know the payoff matrix and must explore the environment to determine rewards associated with variable settings. Second, agents may need to maximize total accumulated reward rather than instantaneous final reward. Third, limited time horizons disallow exhaustive exploration of the environment. We propose and implement a set of novel algorithms that combine decision-theoretic exploration approaches with DCOP-mandated coordination. In addition to simulation results, we implement these algorithms on robots, deploying DCOPs on a distributed mobile sensor network.",
author = "Manish Jain and Matthew Taylor and Milind Tambe and Makoto Yokoo",
year = "2009",
language = "English",
isbn = "9781577354260",
pages = "181--186",
booktitle = "IJCAI-09 - Proceedings of the 21st International Joint Conference on Artificial Intelligence",

}

TY - GEN

T1 - DCOPs meet the realworld

T2 - Exploring unknown reward matrices with applications to mobile sensor networks

AU - Jain, Manish

AU - Taylor, Matthew

AU - Tambe, Milind

AU - Yokoo, Makoto

PY - 2009

Y1 - 2009

N2 - Buoyed by recent successes in the area of distributed constraint optimization problems (DCOPs), this paper addresses challenges faced when applying DCOPs to real-world domains. Three fundamental challenges must be addressed for a class of real-world domains, requiring novel DCOP algorithms. First, agents may not know the payoff matrix and must explore the environment to determine rewards associated with variable settings. Second, agents may need to maximize total accumulated reward rather than instantaneous final reward. Third, limited time horizons disallow exhaustive exploration of the environment. We propose and implement a set of novel algorithms that combine decision-theoretic exploration approaches with DCOP-mandated coordination. In addition to simulation results, we implement these algorithms on robots, deploying DCOPs on a distributed mobile sensor network.

AB - Buoyed by recent successes in the area of distributed constraint optimization problems (DCOPs), this paper addresses challenges faced when applying DCOPs to real-world domains. Three fundamental challenges must be addressed for a class of real-world domains, requiring novel DCOP algorithms. First, agents may not know the payoff matrix and must explore the environment to determine rewards associated with variable settings. Second, agents may need to maximize total accumulated reward rather than instantaneous final reward. Third, limited time horizons disallow exhaustive exploration of the environment. We propose and implement a set of novel algorithms that combine decision-theoretic exploration approaches with DCOP-mandated coordination. In addition to simulation results, we implement these algorithms on robots, deploying DCOPs on a distributed mobile sensor network.

UR - http://www.scopus.com/inward/record.url?scp=77957874916&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=77957874916&partnerID=8YFLogxK

M3 - Conference contribution

SN - 9781577354260

SP - 181

EP - 186

BT - IJCAI-09 - Proceedings of the 21st International Joint Conference on Artificial Intelligence

ER -