WO2017024939A1 - Offloading probabilistic computations in data analytics applications - Google Patents

Offloading probabilistic computations in data analytics applications Download PDF

Info

Publication number
WO2017024939A1
WO2017024939A1 PCT/CN2016/091776 CN2016091776W WO2017024939A1 WO 2017024939 A1 WO2017024939 A1 WO 2017024939A1 CN 2016091776 W CN2016091776 W CN 2016091776W WO 2017024939 A1 WO2017024939 A1 WO 2017024939A1
Authority
WO
WIPO (PCT)
Prior art keywords
call
virtual machine
probabilistic
processor
data
Prior art date
Application number
PCT/CN2016/091776
Other languages
French (fr)
Inventor
Vineet CHADHA
Gopinath Palani
Guangyu Shi
Original Assignee
Huawei Technologies Co., Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co., Ltd. filed Critical Huawei Technologies Co., Ltd.
Priority to CN201680045160.5A priority Critical patent/CN107851060B/en
Priority to EP16834563.5A priority patent/EP3320444A4/en
Publication of WO2017024939A1 publication Critical patent/WO2017024939A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/18File system types
    • G06F16/182Distributed file systems
    • G06F16/1824Distributed file systems implemented using Network-attached Storage [NAS] architecture
    • G06F16/183Provision of network file services by network file servers, e.g. by using NFS, CIFS
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/18File system types
    • G06F16/188Virtual file systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • G06F2009/45583Memory management, e.g. access or allocation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • G06F9/45558Hypervisor-specific management and integration aspects
    • G06F2009/45595Network integration; Enabling network access in virtual machine instances
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/50Indexing scheme relating to G06F9/50
    • G06F2209/509Offload

Definitions

  • Embodiments of the present invention generally relate to the field of analytic computations. More specifically, embodiments of the present invention relate to distributed probabilistic computations.
  • Many data analytics applications e.g., recommendation engines
  • probabilistic computations through specialized probabilistic data structures (e.g., bloom filtering, linear counting, LogLog algorithms, element cardinality, count-min algorithms) based on very large amounts of data. This typically requires fetching large quantities of data from one or more storage servers and transmitting the data to a computational node for processing.
  • Some approaches cause a large increase in network usage, NAS-side CPU usage, memory usage, and NIC usage for transferring raw data to the node for processing.
  • Other drawbacks include data fetching latency and the consumption of application server resources during computation.
  • Very large data sets are common in the web domain or the data analytics domain. Many models are used to process large scale data such as MapReduce.
  • An example of such large scale data processing is the Hadoop ecosystem which is based on querying large data sets for analytics.
  • memory resources are often a limiting factor.
  • Various algorithms have been explored which attempt a compromise between the amount of memory used and the precision required. Because analytics require an estimate of the result of a query, a variance in the expected values is tolerable and safe if based on the computation model used to calculate the value.
  • An approach to offloading probabilistic computations is realized through lightweight (e.g., low overhead) VMs at the storage end.
  • This approach is enabled through the extended distributed protocol by embedding probabilistic operations (e.g., semantics) , along with pointers to data on the storage device.
  • a storage controller performs multiple functions, including deduplication, provisioning, replication, and tiering.
  • Probabilistic algorithms e.g., classification, clustering and collaborative filtering
  • a translator or parser is used to convert the expression (operations) into C objects. Each operation is executed in a containers which provide sandboxing environment (VMs or process level VMs like zeroVM) . Different Operations are pipelined, where output of each operations is input to next operation.
  • an apparatus for offloading a probabilistic computation to a storage device includes an application server comprising a memory and a processor and coupled to a network-attached storage device configured to create a dedicated process in response to a procedural call to a virtual machine container based on a data request.
  • the processor forwards the data request to the network-attached storage device, programs one or more virtual machines to perform a probabilistic computation based on the procedural call, and directs the probabilistic computation to a first virtual machine of the one or more virtual machines.
  • the request for data is transformed into a modified call using a virtualized lookup call.
  • a method for offloading a computation to a storage device includes registering a process identifier with a processor using a user-level application to establish a channel of communication with the processor and fork an output of a virtual machine to create a new process, transforming a request for data into a modified call using a virtualized lookup call, forwarding the modified call to the user-level application using a virtualized NFS procedural call and the channel of communication, creating an inter-process communication channel between a kernel address space of the processor and a user address space, and using the communication channel to forward an NFS call to the virtual machine to perform the probabilistic computation.
  • Figure 1 is a block diagram depicting an exemplary system architecture according to embodiments of the present disclosure.
  • Figure 2 is a block diagram depicting an exemplary configuration for offloading an operation from an application server according to embodiments of the present disclosure.
  • Figure 3 is a block diagram depicting an exemplary probabilistic computation according to embodiments of the present disclosure.
  • Figure 4 is a flowchart depicting an exemplary computer implemented sequence of steps for integrating the VM at the storage end with the application server, according to some embodiments of the present disclosure.
  • Performing probabilistic computations at the storage appliance side may reduce or eliminate network usage, NAS-side CPU utilization, memory usage, and NIC usage necessary for transferring raw data from one or more file servers to a computational node for processing.
  • a methodology to compute the probabilistic computation at a distributed node using a modified application API is disclosed herein.
  • Count-min is an example of a probabilistic data structure used to compute the top k elements stored in data sets. Note that it may require multiple hashing to store and enable querying the state of data sets. The results are often governed by the number of elements and distinct elements in the data sets.
  • another example of a probabilistic data structure is a bloom filter. A bloom filter tracks the state of items through bitsets using several independent hash functions.
  • An example application use-case is a recommendation engine used in popular social networking websites where bloom filtering is used to eliminate a number of candidate items.
  • the query is transferred over a distributed file system protocol and initiates probabilistic computation on preconfigured virtual machines ( “VMs” ) specialized to execute a probabilistic computation algorithm.
  • VMs virtual machines
  • the VMs are configured to receive the input from the input channel, execute the probabilistic algorithm, and send results through the output channel.
  • Such an approach also allows forking (e.g., creating a new processor) or distributing computations among multiple VMs to execute the different algorithms (e.g., parallel computing) to improve the performance and/or accuracy of the computations.
  • the described approach is scalable due to the fact that more or less VMs may be utilized depending on the circumstances.
  • data analytics go through multiple phases (e.g., classification, clustering and collaborative filtering) .
  • Classification attempts to categorize the information based on certain characteristics.
  • Clustering is performed to find objects of similar characteristics or to determine overlap of the object.
  • Collaborative filtering is a technique often used to filter out objects based on multiple data sources.
  • Each phase may have its own computational requirements, and classifying very large sets of data (e.g., petabytes of data) may require a great amount of processing power.
  • the approach of probabilistic computation is realized through lightweight (e.g., low overhead) VMs at the storage end. This approach is enabled through the extended distributed protocol by embedding probabilistic operations (e.g., semantics) , along with pointers to data on the storage device.
  • a storage controller performs multiple functions, including deduplication, provisioning, replication, and tiering.
  • an exemplary system architecture 100 is depicted according to some embodiments of the present invention.
  • An incoming client request 101 is received at application server 102.
  • Application server 102 executes an NFS/CIFS Client application 103 and is coupled to network-attached storage (NAS) device 105 and NAS Device 107 over network fabric 104.
  • NAS Device 107 includes an NAS Server 109 having one or more storage devices 110 and maintains a lightweight virtualization environment 108.
  • virtualization environment 108 has a plurality of virtual machines (e.g., VC1, VC2, VC3 and VC4) .
  • Probabilistic algorithms may be embedded into the traditional protocols (NFS & CIFS) with the “prob_XXX” extensions.
  • a translator or parser is used to convert the expression (operations) into C objects.
  • Each operation is executed in a container which provides a sandboxing environment (VMs or process level VMs like zeroVM) .
  • VMs or process level VMs like zeroVM VMs or process level VMs like zeroVM
  • Application 201 is executed by Application Server 208 and a procedural call is offloaded using NFS Client 202.
  • NAS Device 209 executing NFS Server 203 receives an initial file-handle fetch as well as modified NFS procedural operations.
  • the NFS Server 203 is coupled to kernel communication module 204.
  • Kernel communication module 204 may be executed on a general purpose processor or a network processor, for example.
  • Kernel communications module 204 is configured to provide indirection of a modified RPC call to the user level VM container.
  • User communications module 205 has a user level application thread 206 configured to receive the request parameter from the kernel to configure the sandbox environment.
  • a user program registers a process ID ( “PID” ) to the kernel to establish the channel of communication with the kernel.
  • Application execution inside the storage components can be applied with template 207 to change the behavior of storage containers based on configuration parameters.
  • Network file system is a network protocol based on remote procedural calls ( “RPC” ) .
  • the NFS protocol divides a remote file into blocks of equal size and supports on-demand, block-based transfer of file contents. Multiple procedural calls are defined for NFS protocol such as read, lookup, readdir, remove, etc. Although other file system protocols may be used with the present invention, NFS is a preferred protocol due to its use of remote procedural calls ( “RPC” ) .
  • file or dir objects are addressed using opaque file handles. Any read call is preceded with a lookup to locate the file object to be read. Then the read call is invoked iteratively, where the number of iterations is based on the NFS configuration and file size to be fetched.
  • certain operations e.g., a read operation over a distributed file system
  • a join operation may be seamlessly executed on the backend device and return results to the application in server address space in many cases.
  • Embedding the operation (e.g., the logic or semantics) behind the fetching of data from the backend device into the traditional distributed protocol may reduce or eliminate the cost of fetching large scale data into the server.
  • Process level VMs are generally used to provide either a sandboxed environment for a group of processes to execute securely or to provide platform independence to user applications.
  • Process level virtual machines are a prominent component of virtualized containers such as OpenVZ and Linux container (LXC) .
  • LXC Linux container
  • One aspect of the present invention provides specialized stacks for application execution.
  • One such approach of providing specialized stacks for the operating system is called Exokernel (also known as libOS) .
  • Exokernel reduces the abstraction typically offered by monolithic operating systems ( “O/S” ) .
  • O/S Functionality is exported at the user level and an application may be linked dynamically with specialized O/S runtimes and specialized device drivers to reduce or eliminate the overhead incurred in a traditional operating system.
  • libOS provides interfaces such as process management, paging, file system, and system call interfaces.
  • process management paging
  • file system file system
  • system call interfaces interfaces
  • ZeroVM An example of lightweight virtualization based on libOS principals is ZeroVM.
  • ZeroVM enables isolation of a single process or task into separate containers. The approach allows for the isolation (e.g., sandboxing) of essential tasks only rather than the complete software stack.
  • Application execution inside the storage components can be applied with a template manifest to change the behavior of storage containers based on configuration parameters.
  • various abstraction layers are added on the top of storage appliance to provide customized services for client requests. Linux LUN or volume, for example, may be used to provide the abstraction of capacity.
  • Multiple categories may be used for the template containers for deployment of process level VM on the storage side, such as capacity, throughput, virtualization and functional transformation.
  • lightweight virtualization based on ZeroVM approach is used because the Google (for example) native client is a well-tested approach to providing secured execution for an application and the hypervisor takes less memory to execute the VMs.
  • the User Application API may be modified to include the probabilistic operation of new proposed NFS procedural call prob-lookup.
  • a prob-lookup call invokes a lookup over multiple files, and a probabilistic operation is invoked on top of those file reads in a secured environment (process levelVM) . It is an iterative procedural call which involves looking up the file name, invoking read, and finally applying probabilistic operation computation (e.g., count-min, bloom filter, linear counting, LogLog counting algorithm, etc. ) on the top of it.
  • a system call carries semantics information and a list of files to NFS prob-lookup procedural call.
  • a probabilistic operation may be defined as:
  • a hypervisor can redirect the computation to preconfigured VMs with probabilistic algorithms listening for a request from the client.
  • an exemplary probabilistic computation is depicted according to some embodiments of the present invention.
  • the implementation features DFS Client 303 and DFS Server 306.
  • An incoming client 302 request is received by DFS Client 303 in a first NAS address space 301.
  • the request is forwarded over extended protocol 304 to DFS Server 306 having a second NAS address space 305.
  • Each VM in virtual machine environment 307 e.g. VC1, VC2, VC3 and VCn
  • each VM is configured to execute different probabilistic algorithms (e.g., count-min, LogLog, bloom filter) .
  • a bloom filter may use a space efficient probabilistic data structure. Multiple hashing functions are applied to each element and state data is maintained.
  • a bloom filter is useful for membership queries.
  • a count-min operation is efficient for frequency estimation to determine the top k elements in large data sets, for example.
  • the NFS/CIFS protocol may be modified to support execution of probabilistic based algorithms by adding the following primitives:
  • ⁇ Input (expr) graph or data file to read data long with operation type (classification, clustering collaborative filtering)
  • ⁇ Input (expr) graph or data file to read data long with operation type (classification, clustering collaborative filtering)
  • prob_collabfilter ( “filename-X” , “filename-y” ) ;
  • a kernel module is configured to provide indirection of a modified RPC call to the user level zeroVM container.
  • the container is configured to run the offloaded operation with input/output parameters and a system configuration file.
  • the kernel module creates an inter-process communication ( “IPC” ) channel between the kernel address space and user address space to forward the NFS call to a virtual machine and/or secured container.
  • a user level application thread is configured to receive the request parameter from the kernel to configure the sandbox environment.
  • a user program registers a process ID to the kernel to establish the channel of communication with kernel.
  • ZeroVM cloud hypervisor is used to run offloaded probabilistic operation.
  • ZeroVM is a process level sandbox which allows execution of a single task in a secured environment.
  • ZeroVM features include: restricted memory access for applications; the ability to run foreign compiled source code; and system call restrictions. Abstraction is limited such that a very small surface is exposed to the threat of attack.
  • ZeroVM has multiple channels (e.g. random read and seq write) .
  • VMs in ZeroVM may be forked through configuration file that stores the representative parameter to bootstrap a secured environment. These parameters are dynamically configured by a user level application module when a modified RPC call is invoked with embedded semantics.
  • an exemplary sequence of steps for integrating the VM at the storage end with the application server, where offloading is performed through kernel-user indirection is depicted according to embodiments of the present invention.
  • the approach could employ user level indirection.
  • the following exemplary sequence of events may occur during offloading and deployment.
  • the communication module is loaded as a layer of indirection to pass the semantics-based procedural call to a process level virtual machine.
  • a modified NFS server is loaded into the O/Sto establish an IPC communication channel user space container at step 402.
  • a user level application program is invoked to register the PID with the kernel module and may also fork a VM on-demand at step 403.
  • step 404 the modified NFS server (mountd and nfsd) is boot-strapped.
  • step 405 a client application is started and the request for data is transformed into a modified call through a virtualized lookup call.
  • step 406 a virtualized NFS procedural call is used to forward the modified call to the user level application which may also fork the VM for data operation offloaded into the secured container. The results are sent back to the user application at step 407.

Abstract

An approach to offloading probabilistic computations is described. An application server comprising a memory and a processor and coupled to a network-attached storage device configured to create a dedicated process in response to a procedural call to a virtual machine container based on a data request is disclosed. The processor forwards the data request to the network-attached storage device, programs one or more virtual machines to perform a probabilistic computation based on the procedural call, and directs the probabilistic computation to a first virtual machine of the one or more virtual machines. The request for data is transformed into a modified call using a virtualized lookup call.

Description

OFFLOADING PROBABILISTIC COMPUTATIONS IN DATA ANALYTICS APPLICATIONS
CROSS REFERENCE TO RELATED CASES:
This application claims priority to US Application 14/821,320, filed on August 7, 2015, and entitled “Offloading Probabilistic Computations in Data Analytics Applications, ” which is hereby incorporated by reference in its entirety.
FIELD:
Embodiments of the present invention generally relate to the field of analytic computations. More specifically, embodiments of the present invention relate to distributed probabilistic computations.
BACKGROUND:
Many data analytics applications (e.g., recommendation engines) require probabilistic computations through specialized probabilistic data structures (e.g., bloom filtering, linear counting, LogLog algorithms, element cardinality, count-min algorithms) based on very large amounts of data. This typically requires fetching large quantities of data from one or more storage servers and transmitting the data to a computational node for processing. Currently there are no viable solutions to offload and execute probabilistic based computations on distributed file storage servers (e.g., NFS/CIFS servers) . Some approaches cause a large increase in network usage, NAS-side CPU usage, memory usage, and NIC usage for transferring raw data to the node for processing. Other drawbacks include data fetching latency and the consumption of application server resources during computation.
Very large data sets are common in the web domain or the data analytics domain. Many models are used to process large scale data such as MapReduce. An example of such large scale data processing is the Hadoop ecosystem which is based on querying large data sets for analytics. For such large scale data, memory resources are often a limiting factor. Various algorithms have been explored which attempt a compromise between the amount of memory used and the precision required. Because analytics  require an estimate of the result of a query, a variance in the expected values is tolerable and safe if based on the computation model used to calculate the value.
SUMMARY:
An approach to offloading probabilistic computations is realized through lightweight (e.g., low overhead) VMs at the storage end. This approach is enabled through the extended distributed protocol by embedding probabilistic operations (e.g., semantics) , along with pointers to data on the storage device. A storage controller performs multiple functions, including deduplication, provisioning, replication, and tiering. Probabilistic algorithms (e.g., classification, clustering and collaborative filtering) may be embedded into the traditional protocols (NFS & CIFS) with the defined extensions. A translator or parser is used to convert the expression (operations) into C objects. Each operation is executed in a containers which provide sandboxing environment (VMs or process level VMs like zeroVM) . Different Operations are pipelined, where output of each operations is input to next operation.
In one embodiment, an apparatus for offloading a probabilistic computation to a storage device, is described. The apparatus includes an application server comprising a memory and a processor and coupled to a network-attached storage device configured to create a dedicated process in response to a procedural call to a virtual machine container based on a data request. The processor forwards the data request to the network-attached storage device, programs one or more virtual machines to perform a probabilistic computation based on the procedural call, and directs the probabilistic computation to a first virtual machine of the one or more virtual machines. The request for data is transformed into a modified call using a virtualized lookup call.
In another embodiment, a method for offloading a computation to a storage device is disclosed. The method includes registering a process identifier with a processor using a user-level application to establish a channel of communication with the processor and fork an output of a virtual machine to create a new process, transforming a request for data into a modified call using a virtualized lookup call, forwarding the modified call to the user-level application using a virtualized NFS procedural call and the channel of communication, creating an inter-process communication channel between a kernel  address space of the processor and a user address space, and using the communication channel to forward an NFS call to the virtual machine to perform the probabilistic computation.
BRIEF DESCRIPTION OF THE DRAWINGS:
The accompanying drawings, which are incorporated in and form a part of this specification, illustrate embodiments of the invention and, together with the description, serve to explain the principles of the invention:
Figure 1 is a block diagram depicting an exemplary system architecture according to embodiments of the present disclosure.
Figure 2 is a block diagram depicting an exemplary configuration for offloading an operation from an application server according to embodiments of the present disclosure.
Figure 3 is a block diagram depicting an exemplary probabilistic computation according to embodiments of the present disclosure.
Figure 4 is a flowchart depicting an exemplary computer implemented sequence of steps for integrating the VM at the storage end with the application server, according to some embodiments of the present disclosure.
DETAILED DESCRIPTION:
Reference will now be made in detail to several embodiments. While the subject matter will be described in conjunction with the alternative embodiments, it will be understood that they are not intended to limit the claimed subject matter to these embodiments. On the contrary, the claimed subject matter is intended to cover alternative, modifications, and equivalents, which may be included within the spirit and scope of the claimed subject matter as defined by the appended claims.
Furthermore, in the following detailed description, numerous specific details are set forth in order to provide a thorough understanding of the claimed subject matter. However, it will be recognized by one skilled in the art that embodiments may be practiced without these specific details or with equivalents thereof. In other instances,  well-known methods, procedures, components, and circuits have not been described in detail as not to unnecessarily obscure aspects and features of the subject matter.
Portions of the detailed description that follows are presented and discussed in terms of a method. Embodiments are well suited to performing various other steps or variations of the steps recited in the flowchart of the figures herein, and in a sequence other than that depicted and described herein.
Some portions of the detailed description are presented in terms of procedures, steps, logic blocks, processing, and other symbolic representations of operations on data bits that can be performed on computer memory. These descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. A procedure, computer-executed step, logic block, process, etc., is here, and generally, conceived to be a self-consistent sequence of steps or instructions leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated in a cellular antenna array. It has proven convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like.
It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the following discussions, it is appreciated that throughout, discussions utilizing terms such as “accessing, ” “writing, ” “including, ” “storing, ” “transmitting, ” “traversing, ” “associating, ” “identifying” or the like, refer to the action and processes of a computer, network device, or other electronic computing device that manipulates and transforms data represented as physical (electronic) quantities within the system’s registers and memories into other data similarly represented as physical quantities within the system memories or registers or other such information storage, transmission or display devices.
OFFLOADING PROBABILISTIC COMPUTATIONS IN DATA ANALYTICS APPLICATIONS
As storage operation costs become cheaper, offloading large scale computations to be performed at the storage appliance side (rather than fetching the data into a computing node’s memory to perform the computations) has become more feasible. Performing probabilistic computations at the storage appliance side may reduce or eliminate network usage, NAS-side CPU utilization, memory usage, and NIC usage necessary for transferring raw data from one or more file servers to a computational node for processing. A methodology to compute the probabilistic computation at a distributed node using a modified application API is disclosed herein.
To perform probabilistic computation on a large set of integers, for example, raw data needs can be represented in a form suitable to query for data analytics. An example query would attempt to determine the top k most frequent elements. However, to maintain state information about millions of integers would require a significant amount of memory. Instead, approximate data structures may be used for the analysis. Count-min is an example of a probabilistic data structure used to compute the top k elements stored in data sets. Note that it may require multiple hashing to store and enable querying the state of data sets. The results are often governed by the number of elements and distinct elements in the data sets. Similarly, another example of a probabilistic data structure is a bloom filter. A bloom filter tracks the state of items through bitsets using several independent hash functions. An example application use-case is a recommendation engine used in popular social networking websites where bloom filtering is used to eliminate a number of candidate items.
Probabilistic computation of the data sets using light weight virtualization is described herein. In one approach, the query is transferred over a distributed file system protocol and initiates probabilistic computation on preconfigured virtual machines ( “VMs” ) specialized to execute a probabilistic computation algorithm. The VMs are configured to receive the input from the input channel, execute the probabilistic algorithm, and send results through the output channel. Such an approach also allows forking (e.g., creating a new processor) or distributing computations among multiple VMs to execute the different algorithms (e.g., parallel computing) to improve the performance and/or accuracy of the computations. Furthermore, the described approach  is scalable due to the fact that more or less VMs may be utilized depending on the circumstances.
Typically, data analytics (or machine learning based analytics) go through multiple phases (e.g., classification, clustering and collaborative filtering) . Classification attempts to categorize the information based on certain characteristics. Clustering is performed to find objects of similar characteristics or to determine overlap of the object. Collaborative filtering is a technique often used to filter out objects based on multiple data sources. Each phase may have its own computational requirements, and classifying very large sets of data (e.g., petabytes of data) may require a great amount of processing power.
The approach of probabilistic computation is realized through lightweight (e.g., low overhead) VMs at the storage end. This approach is enabled through the extended distributed protocol by embedding probabilistic operations (e.g., semantics) , along with pointers to data on the storage device. A storage controller performs multiple functions, including deduplication, provisioning, replication, and tiering.
With regard to Figure 1, an exemplary system architecture 100 is depicted according to some embodiments of the present invention. An incoming client request 101 is received at application server 102. Application server 102 executes an NFS/CIFS Client application 103 and is coupled to network-attached storage (NAS) device 105 and NAS Device 107 over network fabric 104. NAS Device 107 includes an NAS Server 109 having one or more storage devices 110 and maintains a lightweight virtualization environment 108. As depicted in Figure 1, virtualization environment 108 has a plurality of virtual machines (e.g., VC1, VC2, VC3 and VC4) .
Probabilistic algorithms (e.g., classification, clustering and collaborative filtering) may be embedded into the traditional protocols (NFS & CIFS) with the “prob_XXX” extensions. A translator or parser is used to convert the expression (operations) into C objects. Each operation is executed in a container which provides a sandboxing environment (VMs or process level VMs like zeroVM) . Different Operations are pipelined, where output of each operation is input to next operation. For example, five different operations may be performed:
Op1 – Read graph from FileA
Op2 – Read graph from FileB
Op3 – Perform Collaborative Filtering
Ope4 – Perform count-min algorithm on output of Op3
Op5 – Save the result model to FileC
Table I
With regard to Figure 2, an exemplary configuration for offloading an operation from an application server is depicted according to some embodiments of the present invention. Application 201 is executed by Application Server 208 and a procedural call is offloaded using NFS Client 202. NAS Device 209 executing NFS Server 203 receives an initial file-handle fetch as well as modified NFS procedural operations. The NFS Server 203 is coupled to kernel communication module 204. Kernel communication module 204 may be executed on a general purpose processor or a network processor, for example. Kernel communications module 204 is configured to provide indirection of a modified RPC call to the user level VM container. User communications module 205 has a user level application thread 206 configured to receive the request parameter from the kernel to configure the sandbox environment. A user program registers a process ID ( “PID” ) to the kernel to establish the channel of communication with the kernel. Application execution inside the storage components can be applied with template 207 to change the behavior of storage containers based on configuration parameters.
NFS protocol and configuration
Network file system is a network protocol based on remote procedural calls ( “RPC” ) . The NFS protocol divides a remote file into blocks of equal size and supports on-demand, block-based transfer of file contents. Multiple procedural calls are defined for NFS protocol such as read, lookup, readdir, remove, etc. Although other file system protocols may be used with the present invention, NFS is a preferred protocol due to its use of remote procedural calls ( “RPC” ) .
In context of the NFS protocol, file or dir objects are addressed using opaque file handles. Any read call is preceded with a lookup to locate the file object to be read. Then the read call is invoked iteratively, where the number of iterations is based on the  NFS configuration and file size to be fetched. With the ongoing trend of increasing file sizes, often referred to as Big Data, and with new workloads emerging in the field, certain operations (e.g., a read operation over a distributed file system) have become very resource intensive. For large scale data, a join operation may be seamlessly executed on the backend device and return results to the application in server address space in many cases. Embedding the operation (e.g., the logic or semantics) behind the fetching of data from the backend device into the traditional distributed protocol may reduce or eliminate the cost of fetching large scale data into the server.
Process level Virtual machine
Process level VMs are generally used to provide either a sandboxed environment for a group of processes to execute securely or to provide platform independence to user applications. Process level virtual machines are a prominent component of virtualized containers such as OpenVZ and Linux container (LXC) . One aspect of the present invention provides specialized stacks for application execution. One such approach of providing specialized stacks for the operating system is called Exokernel (also known as libOS) . Exokernel reduces the abstraction typically offered by monolithic operating systems ( “O/S” ) . O/S Functionality is exported at the user level and an application may be linked dynamically with specialized O/S runtimes and specialized device drivers to reduce or eliminate the overhead incurred in a traditional operating system. Typically libOS provides interfaces such as process management, paging, file system, and system call interfaces. With advent of Big Data, it is highly beneficial to move the data closer to the storage components. For example, converged storage with processing and storage controller on the same dies or very close may eliminate bus latency or network latency through offloading computational tasks to the storage components.
An example of lightweight virtualization based on libOS principals is ZeroVM. ZeroVM enables isolation of a single process or task into separate containers. The approach allows for the isolation (e.g., sandboxing) of essential tasks only rather than the complete software stack. Application execution inside the storage components can be applied with a template manifest to change the behavior of storage containers based on configuration parameters. Typically, various abstraction layers are added on the top of storage appliance to provide customized services for client requests. Linux LUN or  volume, for example, may be used to provide the abstraction of capacity. There are also storage appliance specialized for virtualization.
Multiple categories may be used for the template containers for deployment of process level VM on the storage side, such as capacity, throughput, virtualization and functional transformation. According to some embodiments of the present invention, lightweight virtualization based on ZeroVM approach is used because the Google (for example) native client is a well-tested approach to providing secured execution for an application and the hypervisor takes less memory to execute the VMs.
Probabilistic Computation Offloading
The User Application API may be modified to include the probabilistic operation of new proposed NFS procedural call prob-lookup. A prob-lookup call invokes a lookup over multiple files, and a probabilistic operation is invoked on top of those file reads in a secured environment (process levelVM) . It is an iterative procedural call which involves looking up the file name, invoking read, and finally applying probabilistic operation computation (e.g., count-min, bloom filter, linear counting, LogLog counting algorithm, etc. ) on the top of it. A system call carries semantics information and a list of files to NFS prob-lookup procedural call. A probabilistic operation may be defined as:
Figure PCTCN2016091776-appb-000001
Note that the client and the server are synchronized regarding the probabilistic computation that is feasible at the storage end. A hypervisor can redirect the computation to preconfigured VMs with probabilistic algorithms listening for a request from the client.
With regard to Figure 3, an exemplary probabilistic computation is depicted according to some embodiments of the present invention. As depicted, the implementation features DFS Client 303 and DFS Server 306. An incoming client 302 request is received by DFS Client 303 in a first NAS address space 301. The request is forwarded over extended protocol 304 to DFS Server 306 having a second NAS address space 305. Each VM in virtual machine environment 307 (e.g. VC1, VC2, VC3 and VCn) may be invoked dynamically. According to some embodiments, each VM is configured to execute different probabilistic algorithms (e.g., count-min, LogLog, bloom filter) . A  bloom filter may use a space efficient probabilistic data structure. Multiple hashing functions are applied to each element and state data is maintained. A bloom filter is useful for membership queries. A count-min operation is efficient for frequency estimation to determine the top k elements in large data sets, for example.
The NFS/CIFS protocol may be modified to support execution of probabilistic based algorithms by adding the following primitives:
● prob_compute ( “expr” ) ;
○ Input (expr) = graph or data file to read data long with operation type (classification, clustering collaborative filtering)
○ Output = data model or status
● prob_compute_save ( “expr” , ” filename” ) ;
○ Input (expr) = graph or data file to read data long with operation type (classification, clustering collaborative filtering)
○ Input (filename) = data filename to store the result.
○ output = status
● prob_from ( “filename” ) ;
○ used in conjunction with prob_compute and prob_compute_save to load multidimensional array (each dimension represents one series of events) from a data file
○ Input = filename
○ Output = multidimensional array
● prob_classify ( “filename-X” , “filename-y” ) ;
○ used for categorizing the information based on its characteristics and generate a model
○ Input = filename-X source data file
○ Input = filename-y characteristic element data file
○ Output = trained model
● prob_clustering ( “filename-X” , “filename-y” ) ;
○ clusters objects of similar characteristics
○ Input = filename-X source data file
○ Output = 2-dimensional array (data id and cluster ID )
● prob_collabfilter ( “filename-X” , “filename-y” ) ;
○ filters out data based on multiple data sources
○ Input = filename-X source data file
○ Input = filename-y characteristic source data file
○ Output = 2-dimensional array
Table II
User-kernel communication Module
A kernel module is configured to provide indirection of a modified RPC call to the user level zeroVM container. The container is configured to run the offloaded operation with input/output parameters and a system configuration file. The kernel module creates an inter-process communication ( “IPC” ) channel between the kernel address space and user address space to forward the NFS call to a virtual machine and/or secured container. A user level application thread is configured to receive the request parameter from the kernel to configure the sandbox environment. A user program registers a process ID to the kernel to establish the channel of communication with kernel.
ZeroVM lightweight virtualization
According to some embodiments of the present invention, ZeroVM cloud hypervisor is used to run offloaded probabilistic operation. ZeroVM is a process level sandbox which allows execution of a single task in a secured environment. ZeroVM features include: restricted memory access for applications; the ability to run foreign compiled source code; and system call restrictions. Abstraction is limited such that a very small surface is exposed to the threat of attack. For communication, ZeroVM has multiple channels (e.g. random read and seq write) . VMs in ZeroVM may be forked through configuration file that stores the representative parameter to bootstrap a secured environment. These parameters are dynamically configured by a user level application module when a modified RPC call is invoked with embedded semantics.
With regard to Figure 4, an exemplary sequence of steps for integrating the VM at the storage end with the application server, where offloading is performed through kernel-user indirection, is depicted according to embodiments of the present invention. Alternatively, the approach could employ user level indirection. As depicted in flow chart 400, the following exemplary sequence of events may occur during offloading and  deployment. At step 401, the communication module is loaded as a layer of indirection to pass the semantics-based procedural call to a process level virtual machine. A modified NFS server is loaded into the O/Sto establish an IPC communication channel user space container at step 402. A user level application program is invoked to register the PID with the kernel module and may also fork a VM on-demand at step 403. The process continues at step 404, where the modified NFS server (mountd and nfsd) is boot-strapped. At step 405, a client application is started and the request for data is transformed into a modified call through a virtualized lookup call. At step 406, a virtualized NFS procedural call is used to forward the modified call to the user level application which may also fork the VM for data operation offloaded into the secured container. The results are sent back to the user application at step 407.
Embodiments of the present invention are thus described. While the present invention has been described in particular embodiments, it should be appreciated that the present invention should not be construed as limited by such embodiments, but rather construed according to the following claims.

Claims (20)

  1. An apparatus for offloading a probabilistic computation to a storage device, the apparatus comprising:
    an application server comprising a memory and a processor and coupled to a network-attached storage device configured to create a dedicated process in response to a procedural call to a virtual machine container based on a data request,
    wherein the processor forwards the data request to the network-attached storage device, programs one or more virtual machines to perform a probabilistic computation based on the procedural call, and directs the probabilistic computation to a first virtual machine of the one or more virtual machines, and the procedural call is generated using a virtualized lookup call.
  2. The apparatus of Claim 1, wherein the processor registers a process identifier of a user program to establish a channel of communication between the user program and the network-attached storage device.
  3. The apparatus of Claim 1, wherein the processor changes a behavior of one or more storage containers of the one or more virtual machines using a template.
  4. The apparatus of Claim 1, wherein at least one of the virtual machines is configured to execute a probabilistic algorithm.
  5. The apparatus of Claim 1, wherein at least one of the virtual machines performs a LogLog counting operation.
  6. The apparatus of Claim 1, wherein at least one of the virtual machines performs a membership test.
  7. The apparatus of Claim 1, wherein at least one of the virtual machines performs a bloom filter operation.
  8. In a computer storage system, a method for offloading a probabilistic computation to a storage device, the method comprising:
    registering a process identifier with a processor using a user-level application to establish a channel of communication with the processor;
    transforming a request for data into a modified call using a virtualized lookup call;
    forwarding the modified call to the user-level application using a virtualized NFS procedural call and the channel of communication;
    creating an inter-process communication channel between a kernel address space of the processor and a user address space; and
    using the inter-process communication channel to forward an NFS call to a virtual machine to perform the probabilistic computation.
  9. The method of Claim 8, further comprising passing the probabilistic computation to a secured container of the virtual machine.
  10. The method of claim 9, wherein the secured container provides a sandbox
    environment for executing the probabilistic computation.
  11. The method of claim 8, further comprising forking an output of the virtual machine to create a new process.
  12. The method of claim 11, wherein the new process is executed by a second virtual machine.
  13. The method of Claim 8, wherein the virtual machine is configured to perform a LogLog counting operation.
  14. The method of Claim 8, wherein the virtual machine is configured to perform a membership test.
  15. The method of Claim 8, wherein the virtual machine is configured to perform a bloom filter operation.
  16. A computer program product tangibly embodied in a computer-readable storage device and comprising instructions that when executed by a processor perform a method for offloading a probabilistic computation to a storage device, the method comprising:
    registering a process identifier with a first processor using a user-level application to establish a channel of communication with the first processor and fork an output of a virtual machine to create a new process;
    transforming a request for data into a modified call using a virtualized lookup call;
    forwarding the modified call to the user-level application using a virtualized NFS procedural call and the channel of communication;
    creating an inter-process communication channel between a kernel address space of the first processor and a user address space; and
    using the inter-process communication channel to forward an NFS call to the virtual machine to perform the probabilistic computation.
  17. The method of Claim 16, further comprising passing the probabilistic computation to a secured container of the virtual machine.
  18. The method of claim 17, wherein the probabilistic computation is passed to the secured container as a remote procedural call.
  19. The method of claim 17, wherein the secured container provides a sandbox environment for executing the probabilistic computation.
  20. The method of Claim 17, wherein the virtual machine is configured to perform a membership test.
PCT/CN2016/091776 2015-08-07 2016-07-26 Offloading probabilistic computations in data analytics applications WO2017024939A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201680045160.5A CN107851060B (en) 2015-08-07 2016-07-26 Probabilistic computation in offloading data analysis applications
EP16834563.5A EP3320444A4 (en) 2015-08-07 2016-07-26 Offloading probabilistic computations in data analytics applications

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US14/821,320 US9697034B2 (en) 2015-08-07 2015-08-07 Offloading probabilistic computations in data analytics applications
US14/821,320 2015-08-07

Publications (1)

Publication Number Publication Date
WO2017024939A1 true WO2017024939A1 (en) 2017-02-16

Family

ID=57983932

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2016/091776 WO2017024939A1 (en) 2015-08-07 2016-07-26 Offloading probabilistic computations in data analytics applications

Country Status (4)

Country Link
US (1) US9697034B2 (en)
EP (1) EP3320444A4 (en)
CN (1) CN107851060B (en)
WO (1) WO2017024939A1 (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9934395B2 (en) * 2015-09-11 2018-04-03 International Business Machines Corporation Enabling secure big data analytics in the cloud
KR102511451B1 (en) * 2016-11-09 2023-03-17 삼성전자주식회사 Compuitng system for securely executing a secure application in a rich execution environment
US10824623B2 (en) * 2018-02-28 2020-11-03 Vmware, Inc. Efficient time-range queries on databases in distributed computing systems
AU2019252434B2 (en) * 2018-04-11 2024-03-28 Cornell University Method and system for improving software container performance and isolation
US20190361697A1 (en) * 2018-05-22 2019-11-28 Pure Storage, Inc. Automatically creating a data analytics pipeline
US11150817B2 (en) 2019-02-08 2021-10-19 International Business Machines Corporation Integrating kernel-bypass user-level file systems into legacy applications
US10976963B2 (en) 2019-04-15 2021-04-13 International Business Machines Corporation Probabilistically selecting storage units based on latency or throughput in a dispersed storage network
US11010097B2 (en) 2019-09-18 2021-05-18 International Business Machines Corporation Apparatus, systems, and methods for offloading data operations to a storage system
US20230297268A1 (en) * 2022-03-21 2023-09-21 International Business Machines Corporation Processing data in-memory with memory devices having a crossbar array structure

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100235831A1 (en) 2009-03-12 2010-09-16 Arend Erich Dittmer Method for dynamic configuration of virtual machine
US20110138131A1 (en) 2009-12-09 2011-06-09 Scality, S.A. Probabilistic Offload Engine For Distributed Hierarchical Object Storage Devices
CN102109997A (en) * 2009-12-26 2011-06-29 英特尔公司 Accelerating opencl applications by utilizing a virtual opencl device as interface to compute clouds
US20130191827A1 (en) 2012-01-23 2013-07-25 International Business Machines Corporation System and method to reduce memory usage by optimally placing vms in a virtualized data center
CN103595720A (en) * 2013-11-15 2014-02-19 华为技术有限公司 Offloaded data transferring method, device and client
US20140181804A1 (en) * 2012-12-20 2014-06-26 Hitachi, Ltd. Method and apparatus for offloading storage workload
US8782161B2 (en) * 2011-06-30 2014-07-15 Oracle International Corporation Method and system for offloading computation flexibly to a communication adapter
WO2014110137A1 (en) * 2013-01-10 2014-07-17 Pure Storage, Inc. Performing copies in a storage system

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7003507B2 (en) * 2000-09-18 2006-02-21 International Business Machines Corporation Scalable propagation-based methods for call graph construction
US7475199B1 (en) * 2000-10-19 2009-01-06 Emc Corporation Scalable network file system
US7318229B1 (en) * 2003-07-29 2008-01-08 Sun Microsystems, Inc. Method, system, and program for dispatching a method call
US8336046B2 (en) * 2006-12-29 2012-12-18 Intel Corporation Dynamic VM cloning on request from application based on mapping of virtual hardware configuration to the identified physical hardware resources
WO2009070536A1 (en) * 2007-11-30 2009-06-04 Massachusetts Institute Of Technology Method and apparatus for performing probabilistic inference and providing related solution methods
CN104754008B (en) * 2013-12-26 2019-03-08 伊姆西公司 Network storage node, network store system and the device and method for network storage node
CN104156216B (en) * 2014-08-14 2017-11-03 浪潮(北京)电子信息产业有限公司 A kind of memory management system and method towards cloud computing

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100235831A1 (en) 2009-03-12 2010-09-16 Arend Erich Dittmer Method for dynamic configuration of virtual machine
US20110138131A1 (en) 2009-12-09 2011-06-09 Scality, S.A. Probabilistic Offload Engine For Distributed Hierarchical Object Storage Devices
CN102109997A (en) * 2009-12-26 2011-06-29 英特尔公司 Accelerating opencl applications by utilizing a virtual opencl device as interface to compute clouds
US8782161B2 (en) * 2011-06-30 2014-07-15 Oracle International Corporation Method and system for offloading computation flexibly to a communication adapter
US20130191827A1 (en) 2012-01-23 2013-07-25 International Business Machines Corporation System and method to reduce memory usage by optimally placing vms in a virtualized data center
US20140181804A1 (en) * 2012-12-20 2014-06-26 Hitachi, Ltd. Method and apparatus for offloading storage workload
WO2014110137A1 (en) * 2013-01-10 2014-07-17 Pure Storage, Inc. Performing copies in a storage system
CN103595720A (en) * 2013-11-15 2014-02-19 华为技术有限公司 Offloaded data transferring method, device and client

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
DAVID TALBOT, BLOOM MAPS FOR BIG DATA
MARIANNE DURANDPHILIPPE FLAJOLET, LOGLOG COUNTING OF LARGE CARDINALITIES
See also references of EP3320444A4 *
YAO DEZHONG, ENERGY EFFICIENT TASK SCHEDULING IN MOBILE CLOUD COMPUTING

Also Published As

Publication number Publication date
US20170039080A1 (en) 2017-02-09
EP3320444A1 (en) 2018-05-16
CN107851060B (en) 2020-10-09
CN107851060A (en) 2018-03-27
EP3320444A4 (en) 2019-01-09
US9697034B2 (en) 2017-07-04

Similar Documents

Publication Publication Date Title
US9697034B2 (en) Offloading probabilistic computations in data analytics applications
AU2011238676B2 (en) Virtual application extension points
US8897573B2 (en) Virtual machine image access de-duplication
US20140109087A1 (en) Virtual machine provisioning using replicated containers
US10558702B2 (en) Unified storage system for online image searching and offline image analytics
US20150286414A1 (en) Scanning memory for de-duplication using rdma
US9021478B1 (en) Provisioning virtual machines from template by splitting and building index for locating content portions via content-centric network
CN107113231B (en) Offloading graphics-based computing to a backend device
US10747763B2 (en) Efficient multiple aggregation distinct processing
US20150113092A1 (en) Method and apparatus for distributed enterprise data pattern recognition
US10061725B2 (en) Scanning memory for de-duplication using RDMA
US11681653B2 (en) Namespace representation and enhanced browsability for replicated file systems
US9703788B1 (en) Distributed metadata in a high performance computing environment
US11003629B2 (en) Dual layer deduplication for application specific file types in an information processing system
US11030177B1 (en) Selectively scanning portions of a multidimensional index for processing queries
US10372770B1 (en) Cloud-based platform for semantic indexing of web objects
US20230043929A1 (en) Storage snapshots for nested virtual machines
US11748136B2 (en) Event notification support for nested virtual machines
US9509773B2 (en) Array-based computations on a storage device
US20230153155A1 (en) On-demand co-processing resources for quantum computing
US20210357453A1 (en) Query usage based organization for very large databases
US20230205646A1 (en) Efficient Transparent Rollback of File System Consolidation Migrations
US20240028369A1 (en) Method for reusing identifiers to improve performance of open and close input/output requests for client drive redirection
US20230169033A1 (en) Efficient Transparent Switchover of File System Consolidation Migrations
Hong et al. A new efficient resource management framework for iterative mapreduce processing in large-scale data analysis

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16834563

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2016834563

Country of ref document: EP