WO2011054376A1 - Improved caching of digital contents in p2p networks - Google Patents
Improved caching of digital contents in p2p networks Download PDFInfo
- Publication number
- WO2011054376A1 WO2011054376A1 PCT/EP2009/064547 EP2009064547W WO2011054376A1 WO 2011054376 A1 WO2011054376 A1 WO 2011054376A1 EP 2009064547 W EP2009064547 W EP 2009064547W WO 2011054376 A1 WO2011054376 A1 WO 2011054376A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- file
- network
- node
- content
- sub
- Prior art date
Links
- 238000000034 method Methods 0.000 claims abstract description 23
- 230000004044 response Effects 0.000 claims description 40
- 238000004891 communication Methods 0.000 claims description 8
- 230000004807 localization Effects 0.000 description 41
- 238000012545 processing Methods 0.000 description 22
- 238000012546 transfer Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 239000002957 persistent organic pollutant Substances 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 101150106235 ISPS gene Proteins 0.000 description 1
- 241000270295 Serpentes Species 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000001955 cumulated effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000009792 diffusion process Methods 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/10—Protocols in which an application is distributed across nodes in the network
- H04L67/104—Peer-to-peer [P2P] networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/10—Protocols in which an application is distributed across nodes in the network
- H04L67/104—Peer-to-peer [P2P] networks
- H04L67/1061—Peer-to-peer [P2P] networks using node-based peer discovery mechanisms
- H04L67/1065—Discovery involving distributed pre-established resource-based relationships among peers, e.g. based on distributed hash tables [DHT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/10—File systems; File servers
- G06F16/18—File system types
- G06F16/182—Distributed file systems
- G06F16/1834—Distributed file systems implemented based on peer-to-peer networks, e.g. gnutella
- G06F16/1837—Management specially adapted to peer-to-peer storage networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/10—Protocols in which an application is distributed across nodes in the network
- H04L67/104—Peer-to-peer [P2P] networks
- H04L67/1074—Peer-to-peer [P2P] networks for supporting data block transmission mechanisms
- H04L67/1076—Resource dissemination mechanisms or network resource keeping policies for optimal resource availability in the overlay network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/56—Provisioning of proxy services
- H04L67/568—Storing data temporarily at an intermediate stage, e.g. caching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/957—Browsing optimisation, e.g. caching or content distillation
- G06F16/9574—Browsing optimisation, e.g. caching or content distillation of access to content, e.g. by caching
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L63/00—Network architectures or network communication protocols for network security
- H04L63/04—Network architectures or network communication protocols for network security for providing a confidential data exchange among entities communicating through data packet networks
- H04L63/0428—Network architectures or network communication protocols for network security for providing a confidential data exchange among entities communicating through data packet networks wherein the data content is protected, e.g. by encrypting or encapsulating the payload
- H04L63/0435—Network architectures or network communication protocols for network security for providing a confidential data exchange among entities communicating through data packet networks wherein the data content is protected, e.g. by encrypting or encapsulating the payload wherein the sending and receiving network entities apply symmetric encryption, i.e. same key used for encryption and decryption
Definitions
- the solution according to an embodiment of the present invention relates the information and communications technology field. Particularly, such solution concerns an improved method and a system for accessing content files in peer-to- peer based networks.
- a computer network (or, to be more exact, an overlay network) having the so-called peer-to-peer architecture is formed by a plurality of nodes that simultaneously function as both clients and servers toward the other nodes of the network.
- node it is intended any suitably programmed data processing system capable of receiving, transmitting, storing and processing digital data, such as a personal computer, a laptop, a Personal Digital Assistant (PDA), and the like, managed by a suitable software installed thereon.
- PDA Personal Digital Assistant
- P2P network Unlike the computer networks having a client-server architecture, in a computer network arranged according to a peer-to-peer architecture - from now on, briefly referred to as "P2P network" - any node is capable of starting and/or completing a transaction of digital contents toward any other node. Even tough each node of a P2P network is generally different from the other nodes of the network in terms of computational resource, bandwidth and stored data, all the nodes are considered to be equivalent "peers" each one of them having the rights of downloading/uploading digital contents to/from the others peers.
- a P2P network cumulates the different computational resources and the bandwidth of each peer rather than employing a conventional centralized system where a relatively low number of servers are in charge of managing the transactions of digital contents toward the clients.
- P2P networks are presently exploited in different fields of application, since the advantageous features offered by such network architecture may be useful for many purposes.
- the most common one at the present time is surely the file sharing.
- the content files stored in each peer can be easily shared among the other peers of the network, allowing to exchange them in an efficient way by exploiting the computational resources and the bandwidth cumulated from the various peers themselves.
- content file it is intended a wide category of data, comprising audio data, video data, text data, application software or anything data in digital format.
- P2P networks may be also employed for managing transfer of real time data, such as in the media streaming (both video and audio) and in the telephone (e.g., Voice over Internet Protocol) applications.
- one of the main drawbacks affecting the P2P networks actually employed for implementing file sharing services regards the difficulty with which the content files that are quite uncommon (i.e., the files that are shared by a small number of peers only) manage to spread across the network.
- This problem is felt deeply when said uncommon content files are requested by a relatively high number of peers at the same time, originating a bottleneck in the download process of such content files.
- known solutions provide for creating additional copies of said uncommon content files, so as to increase the sources for the downloading thereof from the requesting peers. In this way, the spread of such uncommon content files across the P2P network would be strongly favored.
- Such additional copies may be stored into a set of peers among the peers forming the P2P network, so as to implement a content files "distributed caching" system.
- each peer is different from the others, and thus a generic peer may have not sufficient disk space for storing the copies of a content file of another peer, or it may not desire to offer its disk space for this purpose.
- an intruder may be able to fill the disk space of the various peers which are offering a cache service with useless data.
- legal problems may arise because the owner of the data processing system corresponding to a determined peer would have the duty of verifying whether the content files to be cached from other peers are copyright protected or not.
- the US patent application US2005/0120133 discloses a distributed cache system adapted to be implemented in hybrid P2P network architectures.
- the indexing system is implemented by means of a centralized server, while only the transfer of the content files is performed in a decentralized way.
- the centralized server has a global vision of the whole network, such server may assess which are the content files shared by the various peers that require to be cached, and decide which are the peers that have to offer disk space for the caching. In this way, the risk of wasting disk space for storing useless data is avoided, since the centralized server can be considered as authoritative.
- this solution does not take into account the amount of traffic that is actually involving each peer of the network; therefore, the centralized server may instruct a peer that is already burdened to cache an additional content file. Moreover, this solution is affected by the same legal drawbacks previously mentioned. In addition, such solution is adapted to be implemented in hybrid P2P network architectures only, while the future trends aim to completely decentralized P2P network architectures.
- the US patent application US2007/0150575 discloses a method for increasing the traffic localization of a network by exploiting a distributed cache system.
- the peers that are within a same ISP may be arranged to form a corresponding distributed caching system.
- This solution provides that each peer is responsible for the caching of a determined subset of content files. If a specific content file is not available (i.e., if the peer that is the responsible for that content file does not store a copy thereof), the content file is downloaded from the internet, and transmitted to the node that is responsible for its caching.
- the known applications that run on the data processing systems forming the peers of a P2P network for the management thereof list the shared content files coming from the various peers into an indexing system, which may implemented by means of one or more dedicated servers or by means of a structured or an unstructured network.
- P2P applications list the shared content files coming from the various peers into an indexing system, which may implemented by means of one or more dedicated servers or by means of a structured or an unstructured network.
- the indexing system is queried by such P2P application itself; the indexing system accordingly replies with a response list of the peers which actually own such content files (or at least chunks thereof). Since the use of such indexing systems is a burden for the network traffic, such response list is not exhaustive, but instead it is limited to a prefixed number of results.
- the set of results included in such response list is generated by the indexing system without taking in consideration the localizations of the peers owning the desired content files with respect to the localization of the peer that have made the request.
- a peer may contact all the peers included in the list, and evaluate their mutual distances by observing the mirror sites thereof.
- This is however an inefficient solution, since the mirror sites are determined by the CDN based on a set of criteria wherein the localization is only a parameter among the others. More importantly, said solution suffers from the same drawback affecting the previous solution, i.e., the number of results in the response list that is limited to a prefixed value.
- P4P provider portal for applications
- the ISPs have to provide on demand a wide range of information regarding the network (such as the policy and the cost of the various destinations) to the indexing system.
- a drawback of this solution is given by the requirement of disclosing confidential information (such as the network topology).
- the main drawback of this solution consists in that the indexing system would be loaded with a too high amount of data, which cannot be efficiently processed using the computing power of the presently available data processing systems.
- the P2P network comprises a plurality of sub-networks each one including a respective plurality of nodes capable of storing content files.
- Each sub-network is associated with a corresponding hash table for the indexing of the content files stored in the nodes of the sub-network; said hash table is distributed among at least some of the nodes of the sub-network.
- the method comprises, for each content file stored in at least one node of a sub-network, storing in the hash table associated with said sub-network a criticality index providing an indication of the degree of sharing of said content file among the nodes of said sub-network.
- the method further comprises having a node caching at least one content file to be shared by other nodes of the sub-network. Said at least one content file is not requested by said node for the fruition thereof; said caching at least one content file is based on the criticality index of said at least one content file.
- Another aspect of the present invention regards a P2P network.
- a still further aspect of the present invention regards a client software executable by a node of a P2P network.
- Figure 1 illustrates an exemplary topology of a P2P network according to an embodiment of the present invention
- Figure 2 depicts a flow chart illustrating the main phases performed by a generic node of the P2P network of Figure 1 during the connection thereto according to an embodiment of the present invention
- Figure 3 schematically illustrates a portion of a distributed hash table according to an embodiment of the present invention
- Figure 4 depicts a flow chart illustrating the main phases performed by a generic node of the P2P network of Figure 1 during look-up operations directed to the search and identification of a specific content file according to an embodiment of the present invention
- Figure 5 schematically illustrates a portion of a distributed hash table according to an embodiment of the present invention
- Figure 6 depicts a flow chart illustrating how a first level table of a distributed hash table is generated according to an embodiment of the present invention
- Figure 7 schematically illustrates a first level table and a second level table of a distributed hash table according to an embodiment of the present invention
- Figure 8 depicts a flow chart illustrating the main phases performed by a node of the P2P network of Figure 1 during a name search and a hash search of a specific content file according to an embodiment of the present invention.
- the P2P network 100 is arranged in two different levels, and namely in a first level, referred to as “legacy network”, and in a second level, referred to as "localized network".
- a first level referred to as "legacy network”
- a second level referred to as "localized network”.
- the main elements of the solutions proposed according to the various embodiments of the present invention reside in the second level of the P2P network 100, i.e., in the localized network.
- the legacy network represents any already known and preexisting P2P network, with which the localized network maintains a backward compatibility.
- the legacy network 105 may be any one among the preexisting global P2P networks presently available, such as for example a decentralized P2P network.
- the legacy network 105 comprises a plurality of nodes which may be arranged according to any one among the known P2P network architectures.
- the legacy network 105 is provided with an indexing system (not shown in the figure), which has the purpose of listing the content files shared by the nodes belonging to the legacy network 105 as well as those stored by nodes forming the localized network.
- indexing system is exploited for performing look-up operations necessary for the identification of which nodes are storing (chunks of) a specific requested content file.
- the legacy network 105 and its corresponding indexing system may be either of the unstructured type or of the structured type, or it may be server based.
- the indexing system may be based on a distributed hash table, such as in the kademlia communication protocol.
- a hash table (or hash map) is a data structure that uses a hash function to efficiently map certain identifiers or keys to associated values.
- the hash function is used to transform the key into the index (the hash) of an array element where the corresponding value is to be sought.
- Distributed hash tables are a class of decentralized distributed systems that provide a lookup service similar to a hash table. Particularly, in a network having associated therewith a distributed hash table, such distributed hash table is adapted to store (key, value) pairs, and any participating node of the network can efficiently retrieve the value associated with a given key.
- Responsibility for maintaining the mapping from keys to values is distributed among the various nodes of the network, in such a way that a change in the set of participants causes a minimal amount of disruption.
- the legacy network 105 will not be described in greater detail later on, because not relevant to the scope of the present invention.
- all the nodes Pj(i) included in a same sub-network 110(i) belong to a same corresponding autonomous system AS(i).
- IP Internet Protocol
- Each sub-network 110(i) forms an independent decentralized P2P network provided with a corresponding indexing system to be exploited for performing lookup operations directed to the search and identification of content files shared among the nodes Pj(i) of the sub-network 110(i).
- the sub-networks 110(i) are of the structured type, with the indexing system of each sub-network 110(i) that is a decentralized distributed system based on a respective distributed hash table DHT(i), like in the kademlia communication protocol. From now on, the indexing system of a generic subnetwork 110(i) will be referred with the term "distributed hash table DHT(i)".
- the distributed hash table DHT(i) is distributed among the nodes Pj(i) forming the corresponding sub-network 110(i); as it is well known to those skilled in the art, this means that the responsibility for managing and maintaining the indexing system is distributed among the nodes Pj(i) of the sub-network 110(i).
- the distributed hash table DHT(i) of a generic sub-network 110(i) provides a look-up service directed to the content files shared among the nodes Pj(i) of such sub-network 100(i) only.
- the nodes Pj(i) of a generic sub-network 110(i) belong to the legacy network 105, too, and are capable of interacting with other nodes of the legacy network 105 that do not belong to such sub-network 110(i) for uploading/downloading content files stored only therein.
- the nodes Pj(i) of the sub-networks 110(i) are back compatible with the other nodes of the legacy network 105, so as to allow nodes Pj(i) of a subnetwork 110(i) which do not store a requested content file to carry out look-up operations directed to the searching of such content file exploiting the indexing system of the legacy network 105.
- the distributed hash table DHT(i) of each sub-network 110(i) is provided with data regarding the localization of the nodes Pj(i) forming such sub-network 110(i); from now on, the data relating the localization of a generic node Pj(i) will be briefly referred to as localization data LD.
- localization data LD the localization data LD of a generic node Pj(i) can be derived from information provided by the ISP by means of which such node Pj(i) has gained access to the P2P network 100.
- the localization data LD of a generic node Pj(i) include three parameters, and namely a first parameter AS which identifies the autonomous system AS(i) corresponding to the sub-network 110(i) including such node Pj(i), a second parameter G which identifies the geographic area in which such node Pj(i) is located, and a third parameter POP which identifies the Point Of Presence (POP) provided by the ISP and used by such node Pj(i) for accessing the P2P network 100.
- a first parameter AS which identifies the autonomous system AS(i) corresponding to the sub-network 110(i) including such node Pj(i) including such node Pj(i)
- a second parameter G which identifies the geographic area in which such node Pj(i) is located
- POP Point Of Presence
- Such three- parameter arrangement is particularly suited for the currently available network topologies, and especially for the European ones; however, similar considerations apply in case of localization data LD including a different number of parameters, e.g. , identifying the state and/or the region in which a node Pj(i) is located.
- the localization data LD have a granularity that corresponds to the POP parameter, which is the more specific one among the three parameters.
- the localization data LD is a string of bits having the same length of two IPv4 addresses. This allows the localization data LD to be transmitted in an easy way by reusing well-known transmission protocols.
- the way the localization data LD are generated and communicated to the distributed hash table DHT(i) of each sub-network 110(i) strongly depends on the technology employed by the various ISP used by the nodes Pj(i) for accessing the network.
- DHCP Dynamic Configuration Protocol
- ISP may transmit the localization data LD to a corresponding node Pj(i) through a specific DHCP option. Since the localization data LD have a granularity corresponding to the parameter POP, it is sufficient that said DHCP option includes the localization data LD corresponding to the POP used by such node Pj(i). This solution does not violate the privacy of the user of the node Pj(i).
- an ISP automatically transmits the localization data LD to all the nodes Pj(i) that it serves, and not only to the nodes Pj(i) that use them for exploiting the services offered by the P2P network 100, the ISP cannot get whether an user of a generic node Pj(i) is actually connected to the P2P network 100 or not. Even if such solution is quite cheap by an implementative point of view, the presence of routers, firewalls and/or the adoption of systems different from the DHCP may impede the transferring of the localization data LD from the ISP to the nodes Pj(i).
- the above mentioned solution may be improved by exploiting a Domain Name System (DNS).
- DNS Domain Name System
- a node Pj(i) which is interested to known its localization data LD may forward a specific requests to the ISP, which will respond based on the IP address of the requester.
- the P2P network 100 is provided with a localization data LD auto-detection system.
- the parameter AS can be easily derived from public databases that can be incorporated in the clients, while the parameter G may be set to a default value.
- the parameter POP may be instead derived through a heuristic way, for example by directly asking the user of a data processing system corresponding to a node Pj(i) the town/city wherein such data processing system is physically located.
- Figure 2 depicts a flow chart 200 illustrating the main phases performed by a generic node Pj(i) during its connection to a corresponding sub-network 110(i) of the P2P network 100 according to an exemplary embodiment of the present invention.
- each sub-network 110(i) is an independent decentralized P2P network, which does not include a dedicated node acting as a server, the bootstrapping of such sub-networks is carried out by using some random nodes Pj(i) as "bootstrap nodes".
- the nodes Pj(i) belonging to a particular sub-network 110(i) that operate as bootstrap nodes are used as entry points of such sub-network 110(i).
- the bootstrap nodes of a sub-network 110(i) are provided with a specific bootstrap flag file BOOT(ASi) which identifies the autonomous system AS(i) corresponding to such sub-network 110(i).
- a data processing system ⁇ e.g., a computer
- a data processing system tries to establish a connection with the P2P network 100 for becoming a node Pj(i) of a sub-network 110(i)
- it firstly accesses the legacy network 105 (block 205) using the services offered by an ISP.
- such data processing system acquires its localization data LD, which includes - among the other parameters - the parameter AS identifying an autonomous system AS(i).
- a first look-up operation is performed for searching if it already exists any node Pj(i) that owns the bootstrap flag file BOOT(ASi) relating to the autonomous system AS(i) identified by the parameter AS (block 210).
- the legacy network 105 replies by providing a response list comprising the IP addresses of the nodes Pj(i) that share the requested bootstrap flag file BOOT(ASi), together with the cryptographic hash of the flag file BOOT(ASi) (block 215).
- the nodes Pj(i) which share such bootstrap flag file BOOT(ASi) are the bootstrap nodes of the sub-network 110(i) corresponding to the autonomous system AS(i).
- the data processing system is registered in the distributed hash table DHT(i) as a new bootstrap node Pj(i) of the sub-network 110(i) (block 225); if instead the actual number of bootstrap nodes of the sub-network 110(i) is higher than the threshold TH (exit branch N of block 220), the data processing system is registered in the distributed hash table DHT(i) as a "standard" node Pj(i) of the sub-network 110(i) (block 230).
- the link of such node Pj(i) to the sub-network 110(i) is actually established.
- the node Pj(i) shares in the distributed hash table DHT(i) the list of the content files that it owns associating them with its localization data LD and (block 235).
- the procedure described above requires to be modified in case the ISP used by the data processing system for the connection to the P2P network 100 makes use of a private addressing system through a Network Address Translation (NAT) apparatus.
- NAT Network Address Translation
- the IP address detected by the legacy network 105 would correspond to that provided by the NAT apparatus, and the bootstrap would not be performed correctly.
- the local IP address of a data processing system which has been registered as a bootstrap node Pj(i) in the distributed hash table DHT(i) is communicated in the place of the cryptographic hash of the bootstrap flag file BOOT(ASi). In this way, a new node Pj(i) may establish a connection with the bootstrap nodes Pj(i) using their local addresses.
- a portion of an exemplary and simplified distributed hash table DHT(i) of a generic sub-network 110(i) is schematically illustrated in Figure 3 according to an embodiment of the present invention.
- the distributed hash table DHT(i) illustrated in Figure 3 comprises a number of rows 305(k) which is equal to the total number of the content files FILE(k) shared by the nodes Pj(i) forming the sub-network 110(i) (in the example at issue, six).
- Each row 305(k) comprises three fields, whose references are 310(k), 320(k) and 325(k), respectively.
- the first field 310(k) of each row 305(k) is adapted to store the hash of the content file FILE(k) determined by applying a hash function to such content file FILE(k), referred to as H(FILE(k)); as will be described in greater detail in the following of the description, the hash H(FILE(k)) is used as a key for the look-up of the content file FILE(k) among the content files shared by the nodes Pj(i) of the subnetwork 110(i).
- the second field 320(k) of each row 305(k) is adapted to store a list showing the nodes Pj(i) of the sub-network 110(i) which store a copy of the content file FILE(k) (or a chunk thereof).
- the list stored in the field 320(k) provides the IP address and the port number of such node Pj(i).
- copies of the content file FILE(l) are stored in the nodes P2(i) and P3(i)
- copies of the content file FILE(2) are stored in the nodes P2(i) and P4(i)
- a copy of the content file FILE(3) is stored in the node Pl(i), and so on.
- the third field 325(k) of each row 305(k) is adapted to store a list showing the localization data LD of each node Pj(i) included in the list of the corresponding second field 320(k).
- the distributed hash table DHT(i) is decentralized, in the sense that the data stored therein and the data structures used for the storing of such data are distributed among the various nodes Pj(i) forming the sub-network 110(i).
- the DHT(i) illustrated in Figure 3 is depicted as a centralized table; however, such table has to be intended as it were distributed among the nodes Pj(i) of the sub-network 110(i).
- the data stored in each row 305(k) - or group of rows - of the distributed hash table DHT(i) of Figure 3 may be stored and managed by a corresponding set of nodes Pj(i) of the sub-network 110(i).
- the node Pj(i) which stores the data corresponding to a row 305(k) of the distributed hash table DHT(i) is referred to as the node Pj(i) responsible for the content file FILE(k) corresponding to said row 305(k).
- the node Pl(i) is the node responsible for the content file FILE(l)
- the node P2(i) is the node responsible for the content files FILE(2) and FILE(3)
- the node P3(i) is the node responsible for the content file FILE(4)
- the node P4(i) is the node responsible for the content files FILE(5) and FILE(6).
- Figure 4 depicts a flow chart 400 illustrating the main phases performed by a generic node Pj(i) connected to the distributed hash table DHT(i) of a sub-network 110(i) during the look-up operations directed to the search and identification of a specific content file FILE(k) to be downloaded according to an embodiment of the present invention.
- node Pj(i) belonging to the sub-network 110(i) desires to find a specific content file FILE(k)
- such node Pj(i) contacts the distributed hash table DHT(i) of the sub-network 110(i) for performing a look-up of such content file
- the look-up operations performed in a DHT network of the type of the P2P network 100 are typically constituted by two different look-up operations to be performed in succession one after the other.
- the first look-up operation also referred to as "name search”
- the result of such first look-up operation is a list comprising the hash of the files shared by the nodes of the network that (at least partially) correspond to the requested file name.
- a second look-up operation is performed, which is also referred to as "hash search”.
- the second look-up operation uses the hash of such file as key; the result of the second look-up operation is a list comprising the IP addresses of the nodes of the network that actually share the file corresponding to such hash.
- the exemplary look-up operation described in the flow chart 400 of Figure 4 makes reference mainly to the hash search. However, a more detailed description of the name search will be provided in the following of the present specification.
- the look-up of the desired content file FILE(k) is started by providing to the distributed hash table DHT(i) a search key (i.e., the name of the content file FILE(k)) together with the localization data LD of the node Pj(i) which has requested the content file (block 410).
- a search key i.e., the name of the content file FILE(k)
- the distributed hash table DHT(i) outputs a corresponding response list - an example of which is illustrated in Figure 4 with the reference 415 - comprising the IP addresses and ports of the nodes Pj(i) belonging to the sub-network 110(i) that actually store copies of the desired file FILE(k) (block 420).
- the results shown in the response list 415 are ordered according to a closeness criterion which quantifies the closeness of the node Pj(i) requesting the content file FILE(k) with the nodes Pj(i) that are listed in the response list 415.
- a closeness criterion consists of comparing the localization data LD sent by the requesting node Pj(i) with the localization data LD corresponding to the nodes Pj(i) that are listed in the response list 415.
- the DHT(i) selects the row 305(k) whose field 310(k) is storing the hash H(FILE(k)) of the requested content file FILE(k), so as to retrieve the list stored in the field 320(k) that shows the nodes Pj(i) of the sub-network 110(i) storing a copy of the requested content file FILE(k) (see Figure 3).
- the response list 415 will include the IP addresses and ports of the nodes Pj(i) listed in the field 320(5), i.e., the nodes Pl(i), P3(i), and P4(i).
- the results provided by the response list 415 are ordered in a sequence based on a comparison between the localization data LD of the requesting node Pj(i) (which data are received together with the search key) and the localization data LD of the nodes Pj(i) listed in the field 320(k) of the row 305(k) corresponding to the requested file FILE(k) (which data are listed in the field 325(k)).
- such sequence is ordered by firstly positioning a first set including the IP addresses and ports of the nodes Pj(i) having both the parameter G and the parameter POP equal to the parameter G and to the parameter POP of the requesting node Pj(i), respectively. Then, such first set is followed in the sequence by a second set including the IP addresses and ports of the nodes Pj(i) having only the parameter G equal to the parameter G of the requesting node Pj(i). The sequence ends with a third set including the IP addressed and ports of the remaining nodes Pj(i) (which however have the same parameter AS, being connected to the same sub-network 110(i)). The order of the nodes Pj(i) within each set is irrelevant.
- the data provided by the distributed hash table DHT(i) to the requesting node Pj(i) through the response list 415 do not explicitly include the localization data LD of the nodes Pj(i) listed therein, in such a way not to violate the privacy of the users of such nodes.
- the requesting node Pj(i) tries to establish a connection with the nodes Pj(i) listed in the response list 415 following the specific order provided in such list (block 430).
- the former node provides its localization data LD to the latter node.
- the download of the requested content file FILE(k) is started (block 435).
- the response list 415 has a number of entries lower than the threshold THR (exit branch Y of block 425), it means that the requested content file FILE(k) is shared by a too low number of the nodes Pj(i) of the sub-network 110(i).
- the request for such content FILE(k) may be forwarded to the legacy network 105 (block 440).
- the nodes Pj(i) of the subnetworks 110(i) are back compatible with the legacy network 105; in this way, lookup operations directed to the searching of such content file FILE(k) may be carried out by exploiting the indexing system of the legacy network 105. The way the legacy network 105 performs such look-up operations will not be described, because not relevant to the scope of the present invention.
- a severe drawback affecting the P2P networks employed for implementing file sharing services regards the difficulty with which the content files FILE(K) that are quite uncommon in a sub-network 110(i) - i.e., those files that are shared only by a small number of nodes Pj(i) of the sub-network 110(i) - can be downloaded.
- This problem is particularly exacerbated when such content files FILE(k) are requested by a relatively high number of nodes Pj(i) of the sub-network 110(i) at the same time.
- a solution according to an embodiment of the present invention provides for a distributed caching system adapted to automatically asses if a generic content file FILE(k) shared by the nodes Pj(i) of a sub-network 110(i) has to be cached or not, and, in the affirmative case, generate additional copies of such content file FILE(k) to be stored in some of the nodes Pj(i) of such sub- network 110(i).
- the distributed hash table DHT(i) of each corresponding sub-network 110(i) stores additional data regarding the actual diffusion of the content files FILE(k) among the nodes Pj(i) of the sub-network 110(i) and the number of pending requests concerning them.
- each row 305(k) of the distributed hash table DHT(i) further includes a fourth field, identified in the figure with the reference 505(k).
- the field 505(k) is adapted to store two numeric parameters Na, Np quantifying how the content file FILE(k) corresponding to the row 305(k) is requested and how it is diffused among the sub-network 110(i).
- the parameter Na corresponds to the number of pending download requests for such content file FILE(k) that have been submitted by nodes Pj(i) of the sub-network 110(i).
- the parameter Na is equal to the number of nodes Pj(i) that have requested the content file FILE(k), but which have not still started the download.
- the parameter Np is instead related to the number of nodes Pi(i) of the sub-network 100(i) actually storing (chunks of) the content file FILE(k).
- the parameter Np corresponding to a specific content file FILE(k) - referred to as Np(FILE(K))- is weighted according to the percentage of the content file FILE(k) actually possessed by the various nodes Pj(i).
- the parameter Np corresponding to a content file FILE(k) may be calculated as:
- Np(FILE(k)) N m + N 15 ⁇ 0,875 + N 50 ⁇ 0,625 + N 25 ⁇ 0,375 + N 0 ⁇ 0,125 , wherein the parameters Nwo, N75, N50, N25, No are the number of nodes Pj(i) of the sub-network 110(i) storing at least the 100%, 75%, 50%, 25% and 0% of the content file FILE(K), respectively.
- the multiplication coefficient of each parameter among the parameters N75, N50, N25, No is obtained by calculating a normalized average value of the percentages corresponding to the same parameter and to the previous parameter.
- the multiplication coefficient 0,875 corresponds to the normalized average value among the percentage 100% and the percentage 75%.
- the concepts of the present invention can be applied even if the parameter Np is calculated in a different way.
- a content file FILE(k) needs to be cached or not based on a comparison between the parameters Na and Np of such content file stored in the distributed hash table DHT(i).
- the higher the parameter Na with respect to the parameter Np the higher the number of pending download requests for such content file FILE(k) with respect to the number of copies actually available within the subnetwork 110(i), and thus, the higher the need of caching such content file FILE(k).
- each row 305(k) of the distributed hash table DHT(i) further includes a fifth field, identified in the figure with the reference 510(k), which is adapted to store a criticality index ICk calculated based on the parameters Na and Np listed in the field 505(k); for example, the criticality index ICk may be determined according to the following relationship:
- t is a predetermined threshold, and 1/8 is a normalizing factor. Therefore, as can be easily inferred by observing the above relationship, the higher the criticality index ICk, the higher the need of caching the content file FILE(k). It has to be appreciated that the criticality index ICk is set to zero if the content file FILE(k) is requested by a small number of nodes Pj(i) (equal or lower than the threshold t).
- each node Pj(i) of a sub-network 110(i) decides (or is instructed) to offer a portion of its disk space for caching purposes, the choice of which content file FILE(k) is to be cached is performed based on the corresponding criticality index ICk. More in detail, each node Pj(i) locally stores a criticality list CLj(i) including the criticality indexes ICk of a set of content files FILE(k) among the content files FILE(k) shared by the nodes of the sub-network 110(i). The choice of which content files FILE(k) are inserted into the criticality list CLj(i) of a node Pj(i) will be described later on.
- a node Pj(i) when a node Pj(i) decides to offer a portion of its disk space for caching purposes, it selects the content files FILE(k) of the criticality list CLj(i) having the highest criticality indexes ICk.
- Each criticality list CLj(i) stored by the node Pl(i) includes a set of rows 515(1) each one corresponding to a specific content file FILE(k).
- the exemplary criticality list CLl(i) illustrated in Figure 5 includes three rows, and particularly a first row 505(1) corresponding to the content file FILE(l), a second row 515(2) corresponding to the content file FILE(5), and a third row 515(3) corresponding to the content file FILE(6).
- Each row 505(1) includes a first field 520(1) adapted to store the hash H(FILE(k)) of the content file FILE(k) corresponding to such row, a second field 525(1) adapted to store the criticality index ICk of the content file FILE(k), and a third field 530(1) adapted to store time data TD(k) providing an indication of the "freshness" of the criticality index ICk, i.e., an indication of when such criticality index ICk has been updated for the last time in the criticality list.
- a node Pj(i) Once a node Pj(i) has chosen the content file FILE(k) to be cached, it starts the download thereof. On this regard, it has to be underlined that the nodes Pj(i) that are downloading a content file FILE(k) for caching purposes are not numerated for the calculation of the parameter Na corresponding to such content file FILE(k).
- the criticality list CLj(i) of a node Pj(i) includes a criticality index ICk which is not still listed in the criticality list CLj(i) of the other node Pj(i)
- the row 515(1) corresponding to such criticality index ICk is copied and added to the criticality list CLj(i) of the latter node Pj(i).
- the criticality list CLj(i) including the oldest version thereof is updated with the criticality index ICk of the other criticality list CLj(i).
- a comparison is made among the time data TD(k) stored in both the critically lists CLj(i).
- the update of the criticality indexes ICk can be advantageously performed during the operation of the P2P network 100, increasing the reliability of the caching service.
- the criticality list CLj(i) of a node Pj(i) that has been recently registered in the distributed hash table DHT(i) may include only the criticality indexes ICk of the content files FILE(k) for which such node is the responsible in the distributed hash table DHT(i), with the passage of time such criticality list CLj(i) is enriched with the information taken from the criticality lists CLj(i) of the other nodes Pj(i) of the sub-network 110(i) with which such node enters in communication.
- the P2P network 100 that has been described until now allows to drastically increase the download efficiency of the content files FILE(k) shared among the nodes Pj(i). This increased efficiency is obtained thanks to the improvements in the look-up operations due to the possibility of localizing the nodes Pj(i) within their sub-networks 110(i), and thanks to the presence of an improved caching system which takes into account the criticality of the content files FILE(k).
- each node Pj(i) included in the sub-network (110(i)) is configured to act both as a peer node performing downloads of requested content files FILE(k) which are stored in other nodes Pj(i), and at the same time as a cache node offering disk space to be exploited by other nodes Pj(i) for caching content files FILE(k) which are not requested by said node Pj(i) for the fruition thereof.
- a higher degree of privacy is guaranteed to the users of the sub-network (110(i)).
- the possibility of offering disk space for caching purposes by the nodes Pj(i) of a P2P network introduces a copyright problem which may be considered to as novel in the P2P system field.
- the user of a node Pj(i) which is offering its disk space for caching a copy of a content file FILE(k) would have the duty of verifying whether such content file FILE(k) is copyright protected or not.
- a solution according to an embodiment of the present invention provides for overcoming said drawbacks by making each user of a node Pj(i) which is caching a content file FILE(k) agnostic to such content file FILE(k).
- both the name search and the hash search performed by a generic node Pj(i) of a sub-network 110(i) are carried out exploiting the distributed hash table DHT(i) corresponding to said sub-network 110(i). Therefore, for the sake of clarity, the distributed hash table DHT(i) will be now considered as a two-level table. Particularly, the first level of the distributed hash table DHT(i) - referred to as first level table - is used during the name search, while the second level of the distributed hash table DHT(i) - referred to as second level table - is used during the hash search. It has to be appreciated that the hash table DHT(i) which has been illustrated in the Figures 3 and 5 was in reality a second level table, being used for the hash search.
- the nodes Pj(i) are made agnostic to the content files FILE(k) that they are caching by means of an encryption system used during both the generation of the distributed hash table DHT(i) and during the name and hash searches.
- Figure 6 depicts a flow chart 600 illustrating how a first level table of a distributed hash table DHT(i) is generated according to an embodiment of the present invention.
- the hash H(FILE(k)) of such content file FILE(k) is determined by applying a hash function to the latter (block 605).
- the content file FILE(k) is encrypted with a symmetric key encryption algorithm (e.g. ⁇ LES, Serpent or Twofish) using the previously calculated hash H(FILE(k)) as cryptographic key (block 610).
- a symmetric key encryption algorithm e.g. ⁇ LES, Serpent or Twofish
- the encrypted version of such content file FILE(k) is referred to as "encrypted content file”, and is identified with the reference CFILE(k).
- the next step provides for calculating the hash H(CFILE(k)) of the encrypted content file CFILE(k) using the same hash function used for generating the hash H(FILE(k)) (block 615).
- such data are used for generating the first level table of the distributed hash table DHT(i), an example of which is illustrated in Figure 7 with the reference lDHT(i).
- the example at issue corresponds to the example already considered in the previous of the present description, wherein the sub-network 110(i) is formed by four nodes Pl(i), P2(i), P3(i) and P4(i), while the number of content files FILE(k) shared by such nodes is equal to 6; each content file FILE(k) is univocally identified by a corresponding name string NAME_FILE(k).
- each name string NAME_FILE(k) identifying a respective content file FILE(k) is formed by a single word; however, the concepts of the present invention can be easily generalized to more complex (and realistic) cases, in which each name string NAME_FILE(k) may be formed by a sequence of more than one words.
- the first level table lDHT(i) comprises a number of rows 705(k) which is equal to the total number of the content files FILE(k) shared by the nodes Pj(i) forming the sub-network 110(i) (in the example at issue, six).
- Each row 705(k) comprises four fields, whose references are 710(k), 720(k), 725(k) and 730(k), respectively.
- the first field 710(k) of each row 705(k) is adapted to store the hash of the name string NAME_FILE(k), which hash is calculated by applying the same hash function used for generating the hash H(FILE(k)); the hash of the name string AME FILE (k) is referred to as H(NAME_FILE(k)).
- H(NAME_FILE(k)) is used as a search key for performing the name search.
- the second field 720(k) is adapted to store the hash H(FILE(k)) of the content file FILE(k), while the third field 725(k) is adapted to store the hash H(CFILE(k)) of the encrypted content file CFILE(k).
- the fourth field 730(k) of the row 705(k) is instead adapted to store the name string NAME_FILE(k) of the content file FILE(k).
- the first level table lDHT(i) is depicted in Figure 7 as a centralized table for the sake of simplicity. However, such table is distributed among the nodes Pj(i) of the sub-network 110(i).
- the node Pl(i) is the node responsible for the content file FILE(l)
- the node P2(i) is the node responsible for the content files FILE(2) and FILE(3)
- the node P3(i) is the node responsible for the content file FILE(4)
- the node P4(i) is the node responsible for the content files FILE(5) and FILE(6).
- Figure 7 also illustrates an example of a corresponding second level table of the distributed hash table DHT(i) - referred to as 2DHT(i)- according to an embodiment of the present invention.
- the second level table 2DHT(i) is equal to the distributed hash table DHT(i) which has been already described making reference to Figure 5, with the only difference that the first field 310(k) of each row 305(k) thereof is adapted to store the hash H(CFILE(k)) of the encrypted content file CFILE(k) instead of the hash H(FILE(k)) of the content file FILE(k).
- the flow chart 800 illustrates the main phases performed by a node Pj(i) of a sub-network 110(i) and by the corresponding distributed hash table DHT(i) during a look-up operation (name search and hash search) of a generic content file FILE(k) identified by a name string NAME_FILE(k) and during the download thereof according to an embodiment of the present invention.
- the flow chart 800 depicts the operations performed by the node Pj(i) and the operations performed by the distributed hash table DHT(i) in two dedicated "lanes", identified in the figure with the references 805 and 810, respectively.
- the first step of the look-up operation consists in the name search.
- the node Pj(i) queries the distributed hash table DHT(i), using the hash H(NAME_FILE(k)) of the name string NAME FILE(k) as a search key (lane 805, block 815).
- the distributed hash table DHT(i) provides a first response list 827 including the hash H(FILE(k)), the hash H(CFILE(k)) and the name strings AME FILE (k) taken from the rows 705(k) of the first level table lDHT(i) having the field 710(k) that stores such hash H(NAME_FILE(k)) (lane 810, block 825). It has to be appreciated that in the simplified example at issue, wherein each content file FILE(k) is identified by a respective name string NAME_FILE(k) formed by a single word.
- the search key used by the node Pj(i) may correspond to only a portion of the complete name string NAME_FILE(k); in this case, the first response list 827 will include the hash H(FILE(k)), the hash H(CFILE(k)) and the (complete) name strings NAME FILE(k) taken from all the rows 705(k) of the first level table lDHT(i) having a word of the (complete) name string NAME_FILE(k) that corresponds to such search key.
- the node Pj(i) selects a specific name string NAME_FILE(k) among those included in the first response list 827 (lane 805, block 830), and temporarily memorizes the hash H(FILE(k)) and the hash H(CFILE(k)) corresponding to such name string NAME_FILE(k) (i.e., the ones taken from the same row 705(k) of the first level table lDHT(i) having the field 730(i) storing such selected name string).
- the next step of the look-up operation consists in the hash search, which is substantially equal to that already described in Figure 4, with the only difference that the search key used by the node Pj(i) for querying the distributed hash table DHT(i) is the hash H(CFILE(k)) of the encrypted content file CFILE(k) instead of the hash H(FILE(k)) of the content file FILE(k).
- the distributed hash table DHT(i) is first queried by the node Pj(i) with a search key equal to the hash H(CFILE(k)) corresponding to the selected name string NAME_FILE(k) (lane 805, block 835).
- search key is provided to the distributed hash table DHT(i) together with the localization data LD of the node Pj(i) which has made the request.
- the distributed hash table DHT(i) has received the hash H(CFILE(k)) and the localization data LD, a check is made on the second level table 2DHT(i), for retrieving the row 305(k) thereof having the field 310(k) that stores such hash H(CFILE(k)) (lane 810, block 840).
- the distributed hash table DHT(i) outputs a corresponding second response list 845 comprising the IP addresses and ports of the nodes Pj(i) belonging to the sub-network 110(i) that actually store copies of the encrypted content file CFILE(k) corresponding to the requested content file FILE(k) (lane 810, block 847).
- second response list 845 includes the nodes Pj(i) listed in the field 320(k) of the second level table 2DHT(i) ordered according to the closeness criterion previously described with reference to Figure 4.
- the node Pj(i) which has requested the download of the content file FILE(k) tries to establish a connection with the nodes Pj(i) listed in the second response list 845 following the specific order provided in such list (lane 805, block 850).
- the node Pj(i) is able to decrypt it using the hash H(FILE(k)) of the (plain) content file FILE(k) that has been previously memorized (in the block 830 of the flow chart 800) as a cryptographic key (lane 805, block 860).
- a node Pj(i) In order to perform a download of a desired content file FILE(k), a node Pj(i) has to known both the hash H(CFILE(k)) - for retrieving the nodes Pj(i) storing copies of the encrypted content file CFILE(k) - and the hash H(FILE(k)) - for decrypting the downloaded encrypted content file CFILE(k).
- the only way for a node Pj(i) to know both the hash H(CFILE(k)) and the hash H(FILE(k)) is being a node Pj(i) that has done a name search using the name string NAME_FILE(k).
- a node Pj(i) is not in possession of both such hash H(CFILE(k)), H(FILE(k)), it is not capable of retrieving a plain version of the content file FILE(k). This condition can be advantageously exploited for making agnostic the nodes Pj(i) to the content files FILE(k) that they are caching.
- a node Pj(i) that has accepted to offer its disk space for caching a copy of a high-requested content file FILE(k) actually stores an encrypted version thereof (i.e., the encrypted content file CFILE(k)) without having the decryption key required for retrieving the plain content file FILE(k) (i.e., the hash H(FILE(k))).
- the user of a node Pj(i) cannot retrieve which type of content FILE(k) is caching, overcoming the abovementioned copyrights drawbacks.
- a traffic engineering feature i.e., a mechanism that enables the ISPs to redirect a part of the traffic from heavily loaded or congested links to less used, and therefore more available, ones.
- the P2P network 100 is provided with a "hierarchical caching" feature.
- each sub-network 110(i) of the P2P network is provided with two different types of additional nodes Pj(i) which are directly controlled by the ISPs, referred to as “cache master nodes” and “cache slave nodes”.
- the cache master nodes have a high knowledge of the topology and the status of the network, and are used for improving the traffic routing among the other nodes.
- the cache master nodes are not used for storing content files to be shared among the other nodes, thus they generate a relatively small amount of additional traffic.
- the cache master nodes may for instance be placed in specific network locations (e.g. in main POPs or in the Network Operations Centre).
- the cache slave nodes lack of any information regarding the topology and the status of the network, but are instead used to store content files for caching purposes. Since the cache slave nodes generate a relatively high amount of traffic, they may be located close to the standard nodes of the network, for example within the main
- the cache master nodes perform look-up operations exploiting the distributed hash table DHT(i) in the same way as previously described; however, once a cache master node has obtained from the distributed hash table DHT(i) the list including the ordered sequence of nodes that actually store copies of the desired content file, it does not initiate the download by self, but rather instructs one or more of the cache slave nodes to perform the download.
- the content files stored in the cache slave nodes are registered in the corresponding distributed hash table DHT(i) by the cache master nodes, and not by the cache slave nodes.
- a master cache node In order to be distinguished from the standard nodes of the network, a master cache node have to inform the corresponding distributed hash table DHT(i), for example setting the parameter G to a dedicated value that does not correspond to any known geographic area.
- the response list comprising the ordered sequence of the nodes storing such content file (such as the response list 415 or the second response list 845) which is provided by the distributed hash table DHT(i) may be modified so as to include at least one cache master node if the requested content file is cached.
- the cache master node may be inserted among the results listed in the response list in a position between the positions of the nodes belonging to the same POP of the node that has performed the request and the position of those belonging to its same geographic area.
- the distributed hash table DHT(i) may shift the position of a cache master node within the nodes listed in the response list by few positions in an unpredictable way, in such a way to avoid that the requesting node is capable of detecting information about the address pools allocated to its POP.
- the method according to the various embodiments of the proposed solution may be implemented through a suitable software (referred to as client software) to be installed on the data processing systems forming the nodes.
- client software may be a stand-alone module or a plug-in.
- client software which may be used to implement each embodiment of the invention is structured in a different way, or if additional modules or functions are provided.
- client software may take any form suitable to be used by any data processing system or in connection therewith (for example, within a virtual machine); particularly, the client software may be in the form of external or resident software, firmware, or microcode (either in object code or in source code - for example, to be compiled or interpreted).
- the client software it is possible to provide the client software on any computer-usable medium; the medium can be any element suitable to contain, store, communicate, propagate, or transfer the software.
- the medium may be of the electronic, magnetic, optical, electromagnetic, infrared, or semiconductor type; examples of such medium are fixed disks (where the software can be pre-loaded), removable disks, tapes, cards, wires, fibers, wireless connections, networks, broadcast waves, and the like.
- the solution according to an embodiment of the present invention lends itself to be implemented even with a hardware structure (for example, integrated in a chip of semiconductor material), or with a combination of software and hardware.
Landscapes
- Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Computer Security & Cryptography (AREA)
- Computer Hardware Design (AREA)
- Computing Systems (AREA)
- Computer And Data Communications (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Information Transfer Between Computers (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
BR112012010503-5A BR112012010503B1 (en) | 2009-11-03 | 2009-11-03 | METHOD TO ALLOW SHARING OF CONTENT FILES BETWEEN US OF A NON-HIERARCHICAL NETWORK, AND, NON-HIERARCHICAL NETWORK |
EP09759695.1A EP2497251B1 (en) | 2009-11-03 | 2009-11-03 | Improved caching of digital contents in p2p networks |
US13/505,959 US8838690B2 (en) | 2009-11-03 | 2009-11-03 | Caching of digital contents in P2P networks |
PCT/EP2009/064547 WO2011054376A1 (en) | 2009-11-03 | 2009-11-03 | Improved caching of digital contents in p2p networks |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/EP2009/064547 WO2011054376A1 (en) | 2009-11-03 | 2009-11-03 | Improved caching of digital contents in p2p networks |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011054376A1 true WO2011054376A1 (en) | 2011-05-12 |
Family
ID=41512483
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2009/064547 WO2011054376A1 (en) | 2009-11-03 | 2009-11-03 | Improved caching of digital contents in p2p networks |
Country Status (4)
Country | Link |
---|---|
US (1) | US8838690B2 (en) |
EP (1) | EP2497251B1 (en) |
BR (1) | BR112012010503B1 (en) |
WO (1) | WO2011054376A1 (en) |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110153737A1 (en) * | 2009-12-17 | 2011-06-23 | Chu Thomas P | Method and apparatus for decomposing a peer-to-peer network and using a decomposed peer-to-peer network |
US10084856B2 (en) * | 2009-12-17 | 2018-09-25 | Wsou Investments, Llc | Method and apparatus for locating services within peer-to-peer networks |
US9432452B2 (en) | 2012-09-10 | 2016-08-30 | Joosy, Inc. | Systems and methods for dynamic networked peer-to-peer content distribution |
US9106721B2 (en) * | 2012-10-02 | 2015-08-11 | Nextbit Systems | Application state synchronization across multiple devices |
US9374420B2 (en) * | 2012-12-14 | 2016-06-21 | Microsoft Technology Licensing, Llc | Content source selection in a P2P network |
US9413846B2 (en) | 2012-12-14 | 2016-08-09 | Microsoft Technology Licensing, Llc | Content-acquisition source selection and management |
US9716749B2 (en) | 2012-12-14 | 2017-07-25 | Microsoft Technology Licensing, Llc | Centralized management of a P2P network |
US10391387B2 (en) | 2012-12-14 | 2019-08-27 | Microsoft Technology Licensing, Llc | Presenting digital content item with tiered functionality |
US20140372551A1 (en) * | 2013-06-13 | 2014-12-18 | Rod G. Fleck | Providing storage and security services with a smart personal gateway device |
US20160197831A1 (en) * | 2013-08-16 | 2016-07-07 | Interdigital Patent Holdings, Inc. | Method and apparatus for name resolution in software defined networking |
JP2015072629A (en) * | 2013-10-03 | 2015-04-16 | 富士通株式会社 | Data processing program and data processing method |
US10841393B2 (en) * | 2018-11-12 | 2020-11-17 | Citrix Systems, Inc. | Systems and methods for secure peer-to-peer caching |
CN111182036B (en) * | 2019-12-12 | 2023-07-25 | 腾讯云计算(北京)有限责任公司 | Data distribution method, network construction method, device, equipment and storage medium |
US11076016B1 (en) * | 2020-11-23 | 2021-07-27 | Verizon Patent And Licensing Inc. | Systems and methods for caching content in a wireless communication network |
CN112836008B (en) * | 2021-02-07 | 2023-03-21 | 中国科学院新疆理化技术研究所 | Index establishing method based on decentralized storage data |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030204602A1 (en) * | 2002-04-26 | 2003-10-30 | Hudson Michael D. | Mediated multi-source peer content delivery network architecture |
WO2004027581A2 (en) * | 2002-09-18 | 2004-04-01 | Skyris Networks, Inc. | Systems, methods and programming for routing and indexing globally addressable objects and associated business models |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7188145B2 (en) | 2001-01-12 | 2007-03-06 | Epicrealm Licensing Llc | Method and system for dynamic distributed data caching |
US7627644B2 (en) | 2003-11-28 | 2009-12-01 | Canon Kabushiki Kaisha | Recipient-centred proactive caching in a peer-to-peer system |
-
2009
- 2009-11-03 WO PCT/EP2009/064547 patent/WO2011054376A1/en active Application Filing
- 2009-11-03 EP EP09759695.1A patent/EP2497251B1/en not_active Not-in-force
- 2009-11-03 BR BR112012010503-5A patent/BR112012010503B1/en not_active IP Right Cessation
- 2009-11-03 US US13/505,959 patent/US8838690B2/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030204602A1 (en) * | 2002-04-26 | 2003-10-30 | Hudson Michael D. | Mediated multi-source peer content delivery network architecture |
WO2004027581A2 (en) * | 2002-09-18 | 2004-04-01 | Skyris Networks, Inc. | Systems, methods and programming for routing and indexing globally addressable objects and associated business models |
Non-Patent Citations (3)
Title |
---|
CLARKE I ET AL: "Freenet: a distributed anonymous information storage and retrieval system", PROCEEDINGS OF THE INTERNATIONAL WORKSHOP ON DESIGN ISSUES INANONYMITY AND UNOBSERVABILITY, XX, XX, 1 January 2001 (2001-01-01), pages 1 - 21, XP002974738 * |
DABEK F ET AL: "Wide-area cooperative storage with CFS", PROCEEDINGS OF THE ACM SYMPOSIUM ON OPERATING SYSTEMS PRINCIPLES, XX, XX, 1 October 2001 (2001-10-01), pages 202 - 215, XP002974739 * |
LIHANG YING ET AL: "pcVOD: Internet Peer-to-Peer Video-On-Demand with Storage Caching on Peers", PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON DISTRIBUTEDMULTIMEDIA SYSTEMS, XX, XX, 1 September 2005 (2005-09-01), XP002460362 * |
Also Published As
Publication number | Publication date |
---|---|
EP2497251B1 (en) | 2018-03-07 |
EP2497251A1 (en) | 2012-09-12 |
US20120221646A1 (en) | 2012-08-30 |
BR112012010503A2 (en) | 2016-03-15 |
BR112012010503B1 (en) | 2021-11-16 |
US8838690B2 (en) | 2014-09-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2497250B1 (en) | Sharing of digital contents in p2p networks exploiting localization data | |
EP2497251B1 (en) | Improved caching of digital contents in p2p networks | |
Alimi et al. | Application-layer traffic optimization (ALTO) protocol | |
US8302204B2 (en) | Secure distributed item-level discovery service using secret sharing | |
KR101330392B1 (en) | Network nodes and methods for data authorization in distributed storage networks | |
RU2433461C2 (en) | Interaction between neighbourhoods within federation on rendezvous mechanism | |
US20080130516A1 (en) | P2p Overplay Network Construction Method and Apparatus | |
JP2015165657A (en) | Content name resolution for information oriented networking | |
JP2006174417A (en) | Method and system for associating resource requests with corresponding resources | |
US10187460B2 (en) | Peer-to-peer sharing in a content centric network | |
JP2009277234A (en) | Method for facilitating communication in content centric network | |
JP5666719B2 (en) | Search in peer-to-peer networks | |
US8244867B2 (en) | System and method for the location of caches | |
Ford | UIA: A global connectivity architecture for mobile personal devices | |
Ashraf et al. | Comparative analysis of unstructured P2P file sharing networks | |
WO2020010270A1 (en) | Dynamic routing using a distributed hash table | |
Ngaffo et al. | Information-centric networking challenges and opportunities in service discovery: A survey | |
US20100212006A1 (en) | Peer-to-peer traffic management based on key presence in peer-to-peer data transfers | |
Lee et al. | Advanced node insertion attack with availability falsification in Kademlia-based P2P networks | |
Nandan et al. | GhostShare-reliable and anonymous P2P video distribution | |
Dutta et al. | Naming for Unique Content Identification | |
Cowan | S4h: A Peer-to-Peer Search Engine with Explicit Trust | |
Cojocar | BBUFs: A new lookup mechanism based on IPV6 | |
Ross et al. | P2P Systems | |
Kiesel et al. | RFC 7285: application-layer traffic optimization (ALTO) protocol |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09759695 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13505959 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009759695 Country of ref document: EP |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112012010503 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112012010503 Country of ref document: BR Kind code of ref document: A2 Effective date: 20120503 |