Networking Seminar
Fall 2005
Talks:
|  | Michal Feldman | |
|  | Shay Kutten | |
|  |  |  | 
|  | Yaron De Levie | |
|  |  |  | 
|  | Boaz Patt-Shamir | Trust, Collaboration and Recommendations in Peer-to-Peer Systems | 
|  | Nir Halachmi | Protecting Bursty Applications
  Against Traffic Aggressiveness | 
|  | Shmuel Friedland | |
| Jan    
  5, 2006 | Eyal Even-Dar | |
| Jan  
  12, 2006 | -------- |  | 
| Jan   
  19, 2006 | Zvi Lotker | Publish and Perish: Definition and Analysis of  an
  $n$-Person  Publication Impact Game | 
| Jan   
  26, 2006 | Allon Shafrir | Approximate Top-k Queries in Sensor Networks  | 
| Feb 3, 2006 | Shie Mannor | Asymptotics of Efficiency Loss in
  Competitive Market Mechanisms | 
Michal Feldman, TAU and HU
Title:
Hidden-Action in Multi-Hop Routing
Abstract:
In multi-hop networks, the actions taken by individual intermediate
nodes are typically hidden from the communicating endpoints; all the
endpoints can observe is whether or not the end-to-end transmission
was successful. Therefore, in the absence of incentives to the
contrary, rational (i.e., selfish) intermediate nodes may choose to
forward packets at a low priority or simply not forward packets at
all.  Using a principal-agent model, we show how the hidden-action 
problem can be overcome through appropriate design of contracts, in
both the direct (the endpoints contract with each individual router)
and recursive (each router contracts with the next downstream router)
cases.  We further demonstrate that per-hop monitoring does not 
necessarily improve the utility of the principal or the social welfare
in the system. In addition, we generalize existing mechanisms that
deal with hidden-information to handle scenarios involving both
hidden-information and hidden-action. 
Title: Locality of Fault Tolerance
Abstract:
Traditional fault tolerance methods are global in nature. Some involve
resetting all
the network nodes. Others use time-outs that imply waiting to the slowest and
furthest node, etc.
Local detection of faults is based on the following observation: even 
computations that cannot be performed locally, can be
verified locally.
Hence, it is possible to verify locally (cooperating only with neighbors) 
global correctness predicates. If the predicate does not hold, then measures
for correctness can be invoked.
The cost of correction can also be, sometimes, local to the faults, in the 
sense that when the extent of the faults is lower, the cost of the recovery is 
lower too.
The talk will mention results from several papers by several authors, including
a very recent paper.
Title: Space and Step Complexity Efficient Adaptive Collect
Abstract: Space and step complexity efficient deterministic adaptive
to total contention collect algorithms are presented. One of them has an
optimal O(k) step and O(n) space complexities, but restrict
the processes
identifiers size to O(n). Where n is the total number of processes in the
system and k is the total contention, the total number of processes active
during the execution. Unrestricting the name space increases the space
complexity to O(n^2) leaving the step complexity at
O(k). To date all
deterministic adaptive collect algorithms that we are aware of are either
nearly quadratic in their step complexity or their memory overhead is
exponential in n.
Speaker: Boaz Patt-Shamir, TAU
Title: Trust, Collaboration and Recommendations in Peer-to-Peer Systems 
Abstract:
In peer-to-peer (p2p) systems, work is distributed among the
participating peers, unlike the classical client-server architecture,
where work is done by a central server. One of the fundamental
problems facing p2p systems is that different peers may have different
interests: in extreme cases, some peers may even wish to destroy the
usability of the system. It is therefore necessary to develop a
(possibly implicit) notion of trust, so as to allow peers to continue
functioning in the face of diverse, potentially hostile peers. In this
talk I will describe a line of recent research which studies the
algorithmic aspect of concepts such as "trust," "collaborative
filtering" and "recommendation systems."
Based on various papers with Alon, Awerbuch, Azar, Lotker, Peleg and Tuttle.
Title: Protecting Bursty Applications Against Traffic Aggressiveness
Speaker: Nir Halachmi (IDC)
Time: Thurday, 22/12. 10:30
Location: Shreiber 309
Abstract
Aggressive use of networks, in particular the Internet, either by malicious or innocent users, threats the service availability and quality of polite applications. Common queueing mechanisms which supposedly solve the problem, are shown in this work to be ineffective for bursty applications including Web applications. This can be exploited by malicious users to conduct a new kind of denial of service attacks.
We propose a new traffic control mechanism called {\it Aggressiveness
Protective Queueing (APQ)} which is  based on
attributing importance weights to the users and which solves this problem by
dynamically decreasing the weight of the aggressive users. The actual weight
used for a flow is a dynamically varying parameter reflecting the past bandwidth
usage of the flow. We show that under heavy load (deterministic model), {\it APQ} significantly restricts the amount of traffic an
aggressive user can send and bounds it to at most {\it twice} the amount of
traffic sent by a polite (regular) user. Simulation results 
demonstrate the effectiveness of APQ under a stochastic environment.
Joint work with Anat Bremler-Barr (IDC) and Hanoch Levy (TAU)
Speaker: Shmuel Friedland
Title: The Role of Singular Value Decomposition in Data Analysis
Abstract:
The singular value decomposition (SVD) of an $m\times n$ matrix
emerged as a very important tool in data analysis, data
compression and data recovery in computer science, electrical
engineering and biology.
In this talk we will explain the significance of SVD and some
recent applications to the following topics:
1. Fast low rank approximation of matrices using Monte-Carlo techniques.
 2.  A joint SVD decomposition of two or more
matrices to compare several biological processes.
 3. 
Estimation of missing values in given matrix data using the inverse eigenvalue problems techniques, and  their applications to to DNA microarrays and image
processing.
Most of the results can be found the following recent papers,  which are available
at  http://www.math.uic.edu/~friedlan/research.html
 
Routing Without Regret
 
There has been substantial work developing simple, efficient
 ``no-regret'' algorithms for a number of adaptive game-playing
 problems including online routing.  There has also been
 substantial work on analyzing properties of Nash equilibria
in
 routing games. In this paper, we consider the question: if each
 player in a routing game uses a no-regret strategy, will behavior
 converge to a Nash equilibrium, and under what
conditions and in
 what sense? Note that in general games the answer is known to be
 {\em no} in a strong sense, but routing games
have substantially
 more structure.
 
In this paper we give a number of positive results.  Our main result is
 that in the setting of multicommodity flow and
infinitesimal agents, the
 time-average flow is an $\epsilon$-Nash
equilibrium---a flow $f$ whose
 cost is within $\epsilon$ of the cheapest paths possible given $f$---for
 $\epsilon$ approaching 0 at a rate that depends polynomially
on the
 players' regret bounds and the maximum slope of any latency
 function. Moreover, we show the dependence on slope is necessary.  We
 also give stronger results for special cases such as the case of $n$
 parallel links, and also consider the finite-size (non-infinitesimal)
 load-balancing model of Azar \cite{AZ98}. Our motivations are similar to
 the recent work of Fischer and V\"{o}cking \cite{FV04} and we discuss
 relations to their results as well.
 
Joint work with A. Blum and K. Ligett
Speaker: Zvi Lotker, CWI
Title : Publish and Perish: Definition and Analysis
of  an $n$-Person  Publication Impact Game
We consider the following abstraction of competing publications. 
There are $n$ players vying for the attention of the audience. The 
attention of the audience is abstracted by a single slot which 
holds, at any given time, the name of the latest release. Each 
player needs to choose, ahead of time, when to release its product, 
and the goal is to maximize the amount of time his product is the 
latest release.  Formally, each player $i$
chooses a point 
$x_i\in[0,1]$, and his
payoff is the distance from its point $x_i$ 
to the next larger point, or to $1$ if $x_i$ is the
largest.  For 
this game, we give a complete characterization of the Nash 
equilibrium for the two-player, continuous-action game, and, more 
important, we give an efficient algorithm to compute the symmetric 
Nash equilibrium for the $n$-player, discrete-action game.  In both 
cases, we show that the (symmetric) equilibrium is unique.  Our 
algorithmic approach to the $n$-player game is non-standard in that 
it does not involve solving a system of differential equations. We 
believe that our techniques can be useful in the analysis of other 
timing games.
This is join work with   Boaz Patt-Shamir,
Mark R. Tuttle 
Title: Approximate Top-k Queries in Sensor Networks 
Abstract:
We consider a distributed system where each node keeps a local count for items (similar to elections where nodes are ballots and items are candidates). A top-k query in such a system asks which are the k items whose global count, across all nodes in the system, is the largest. In this paper we present a Monte-Carlo algorithm that outputs, with high probability, a set of k candidates which approximates the top-k items. The algorithm is motivated by sensor networks in that it focuses on reducing the communication complexity. In contrast to previous algorithms, the communication complexity depends only on the global scores and not on the number of nodes or the partition of scores among nodes. If the number of nodes is large, our algorithm can dramatically reduce the communication complexity when compared to deterministic algorithms. The complexity of our algorithm is close to a lower bound on the cell-probe complexity of any non-interactive top-k approximation algorithm. If the global scores are relatively rapidly decreasing (such as the Geometric or Zipf distributions), our algorithm achieves polylog communication complexity per node.
Joint work with Boaz Patt-Shamir.
Title: Asymptotics of Efficiency Loss in Competitive Market
Mechanisms
 
Abstract
 
Decentralized control
mechanisms for networks have the objective of
maximizing social welfare in the face of heterogeneous demand and lack
of coordination among agents. In this talk we consider a probabilistic
setup, where the agents are assumed to be sampled from a given population.
We consider the efficiency loss, in terms of social welfare, between the
best centralized resource allocation mechanism and a simple decentralized
one introduced by Kelly. An algebraic bound on this efficiency loss exists
in the most general setup. We present asymptotic results concerning the
convergence of the loss of efficiency under the random sampling model.
In particular, we show that the loss of efficiency tends to zero
with high probability under some standard assumptions. Moreover, we derive
finite sample bounds for the efficiency loss as a function of the number of
users.
If, however, the assumption of bounded utility functions is relaxed, the loss
of
efficiency no longer converges to zero in a strong probabilistic sense.
These results are further extended to random networks where we analyze the
asymptotic behavior under a similar allocation mechanism.
Collaborating simulations are also presented.