fair division of indivisible goods under risk
play

Fair division of indivisible goods under risk Sylvain Bouveret - PowerPoint PPT Presentation

Fair division of indivisible goods under risk Sylvain Bouveret Charles Lumet Michel Lematre Onera Toulouse Workshop on Social Choice and Artificial Intelligence @ IJCAI 2011 Barcelona, July 16, 2011 Introduction Classical fair division


  1. Fair division of indivisible goods under risk Sylvain Bouveret Charles Lumet Michel Lemaître Onera Toulouse Workshop on Social Choice and Artificial Intelligence @ IJCAI 2011 Barcelona, July 16, 2011

  2. Introduction Classical fair division Fair division of indivisible goods. . . We have : a finite set of objects O = { 1 ,..., l } a finite set of agents A = { 1 ,..., n } having some preferences on the set of objects they may receive We want : an allocation π : A → 2 O such that π i ∩ π j = ∅ if i � = j and which takes into account the agents’ preferences Fair division of indivisible goods under risk 2 / 19 �

  3. Introduction Examples A toy-example : A set of bottles of wine to share. . . Objects : bottles of wine Agents : wine amateurs Fair division of indivisible goods under risk 3 / 19 �

  4. Introduction Examples A toy-example : A set of bottles of wine to share. . . Objects : bottles of wine Agents : wine amateurs A more realistic example : A co-funded Earth-observing satellite to operate. . . Agents : the countries that have co-funded the satellite Objects : observation requests posted by the agents Fair division of indivisible goods under risk 3 / 19 �

  5. Introduction Centralized allocation A classical way to solve the problem : Ask the agents to give a score (weight, utility. . .) w ( o ) to each object o Consider that they have additive preferences → u ( π ) = � o ∈ π w ( o ) Find an allocation that maximizes min i ∈A u ( π ( i )) (egalitarian solution [Rawls, 1971]) Fair division of indivisible goods under risk 4 / 19 �

  6. Introduction Centralized allocation A classical way to solve the problem : Ask the agents to give a score (weight, utility. . .) w ( o ) to each object o Consider that they have additive preferences → u ( π ) = � o ∈ π w ( o ) Find an allocation that maximizes min i ∈A u ( π ( i )) (egalitarian solution [Rawls, 1971]) Example : 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 π = �{ j 1 , j 2 } , { j 3 }� → uc ( π ) = min ( 4 + 5 , 4 ) = 4 π ′ = �{ j 1 } , { j 2 , j 3 }� → uc ( π ′ ) = min ( 5 , 4 + 1 ) = 5 5 4 2 i 1 4 1 4 i 2 Fair division of indivisible goods under risk 4 / 19 �

  7. Introduction Centralized allocation A classical way to solve the problem : Ask the agents to give a score (weight, utility. . .) w ( o ) to each object o Consider that they have additive preferences → u ( π ) = � o ∈ π w ( o ) Find an allocation that maximizes min i ∈A u ( π ( i )) (egalitarian solution [Rawls, 1971]) The Santa-Claus problem [Bansal and Sviridenko, 2006] Bansal, N. and Sviridenko, M. (2006). The santa claus problem. In Proceedings of the thirty-eighth annual ACM symposium on Theory of computing , pages 31–40. ACM. Fair division of indivisible goods under risk 4 / 19 �

  8. Introduction Adding uncertainty Now, we might be unsure of the quality of the objects when they are allocated. Fair division of indivisible goods under risk 5 / 19 �

  9. Introduction Adding uncertainty Now, we might be unsure of the quality of the objects when they are allocated. The bottles can be tainted. The weather can be cloudy over the observed area. Fair division of indivisible goods under risk 5 / 19 �

  10. Introduction Adding uncertainty Now, we might be unsure of the quality of the objects when they are allocated. The bottles can be tainted. The weather can be cloudy over the observed area. If we have some probabilistic information on the quality of an object, how can we take it into account in the allocation process ? Fair division of indivisible goods under risk 5 / 19 �

  11. Introduction Adding uncertainty Now, we might be unsure of the quality of the objects when they are allocated. The bottles can be tainted. The weather can be cloudy over the observed area. If we have some probabilistic information on the quality of an object, how can we take it into account in the allocation process ? We assume that : each object can be in two possible states : good or bad (bad = utility 0) each object o has a probability p ( o ) to be good these probabilities are independent Fair division of indivisible goods under risk 5 / 19 �

  12. The model Resource allocation under risk Resource allocation problem A tuple ( A , O , W , p ) with : A = { 1 ,.., n } a set of agents O = { 1 ,.., l } a set of objects W ∈ M n , l ( R + ) a matrix of weights (given by agents to objects) p ∈ [ 0 , 1 ] l the probability for each object to be in good state. Fair division of indivisible goods under risk 6 / 19 �

  13. The model Resource allocation under risk Resource allocation problem A tuple ( A , O , W , p ) with : A = { 1 ,.., n } a set of agents O = { 1 ,.., l } a set of objects W ∈ M n , l ( R + ) a matrix of weights (given by agents to objects) p ∈ [ 0 , 1 ] l the probability for each object to be in good state. Notations : S : the set of 2 l states of the world good ( s ) ⊆ O : the set of objects in good states in s ∈ S u i , s ( π ) the utility of agent i in s with allocation π . Fair division of indivisible goods under risk 6 / 19 �

  14. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Fair division of indivisible goods under risk 7 / 19 �

  15. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 i 1 5 4 2 i 2 4 1 4 Fair division of indivisible goods under risk 7 / 19 �

  16. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 i 1 5 4 2 i 2 4 1 4 Risk : ∀ j , p j = 0 . 5 Fair division of indivisible goods under risk 7 / 19 �

  17. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 i 1 5 4 2 i 2 4 1 4 Risk : ∀ j , p j = 0 . 5 Allocations : π = �{ j 1 , j 2 } , { j 3 }� π ′ = �{ j 1 } , { j 2 , j 3 }� Fair division of indivisible goods under risk 7 / 19 �

  18. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 i 1 5 4 2 Profiles : i 2 4 1 4 � � 0 0 4 4 5 5 9 9 − → π 0 4 0 4 0 4 0 4 Risk : ∀ j , p j = 0 . 5 � � π ′ − 0 0 0 0 5 5 5 5 → 0 4 1 5 0 4 1 5 Allocations : π = �{ j 1 , j 2 } , { j 3 }� π ′ = �{ j 1 } , { j 2 , j 3 }� Fair division of indivisible goods under risk 7 / 19 �

  19. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 i 1 5 4 2 Profiles : i 2 4 1 4 � � 0 0 4 4 5 5 9 9 − → π 0 4 0 4 0 4 0 4 Risk : ∀ j , p j = 0 . 5 � � π ′ − 0 0 0 0 5 5 5 5 → 0 4 1 5 0 4 1 5 Allocations : π = �{ j 1 , j 2 } , { j 3 }� π ′ = �{ j 1 } , { j 2 , j 3 }� Fair division of indivisible goods under risk 7 / 19 �

  20. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 i 1 5 4 2 Profiles : i 2 4 1 4 � � 0 0 4 4 5 5 9 9 − → π 0 4 0 4 0 4 0 4 Risk : ∀ j , p j = 0 . 5 � � π ′ − 0 0 0 0 5 5 5 5 → 0 4 1 5 0 4 1 5 Allocations : π = �{ j 1 , j 2 } , { j 3 }� π ′ = �{ j 1 } , { j 2 , j 3 }� Fair division of indivisible goods under risk 7 / 19 �

  21. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 i 1 5 4 2 Profiles : i 2 4 1 4 � � 0 0 4 4 5 5 9 9 − → π 0 4 0 4 0 4 0 4 Risk : ∀ j , p j = 0 . 5 � � π ′ − 0 0 0 0 5 5 5 5 → 0 4 1 5 0 4 1 5 Allocations : π = �{ j 1 , j 2 } , { j 3 }� π ′ = �{ j 1 } , { j 2 , j 3 }� Fair division of indivisible goods under risk 7 / 19 �

  22. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 i 1 5 4 2 Profiles : i 2 4 1 4 � � 0 0 4 4 5 5 9 9 − → π 0 4 0 4 0 4 0 4 Risk : ∀ j , p j = 0 . 5 � � π ′ − 0 0 0 0 5 5 5 5 → 0 4 1 5 0 4 1 5 Allocations : π = �{ j 1 , j 2 } , { j 3 }� π ′ = �{ j 1 } , { j 2 , j 3 }� Fair division of indivisible goods under risk 7 / 19 �

  23. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } Preferences : j 1 j 2 j 3 i 1 5 4 2 Profiles : i 2 4 1 4 � � 0 0 4 4 5 5 9 9 − → π 0 4 0 4 0 4 0 4 Risk : ∀ j , p j = 0 . 5 � � π ′ − 0 0 0 0 5 5 5 5 → 0 4 1 5 0 4 1 5 Allocations : π = �{ j 1 , j 2 } , { j 3 }� π ′ = �{ j 1 } , { j 2 , j 3 }� Fair division of indivisible goods under risk 7 / 19 �

  24. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } � � 0 0 4 4 5 5 9 9 0 4 0 4 0 4 0 4 π � � 0 0 0 0 5 5 5 5 0 4 1 5 0 4 1 5 π ′ Fair division of indivisible goods under risk 8 / 19 �

  25. The model Example 3 objects { j 1 , j 2 , j 3 } , 2 agents { i 1 , i 2 } � � � � 0 0 4 4 5 5 9 9 4 . 5 E − → 0 4 0 4 0 4 0 4 2 π � � 0 0 0 0 5 5 5 5 0 4 1 5 0 4 1 5 π ′ Fair division of indivisible goods under risk 8 / 19 �

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend