US20140032527A1 - Parallel Processing of Data - Google Patents

Parallel Processing of Data Download PDF

Info

Publication number
US20140032527A1
US20140032527A1 US14/033,145 US201314033145A US2014032527A1 US 20140032527 A1 US20140032527 A1 US 20140032527A1 US 201314033145 A US201314033145 A US 201314033145A US 2014032527 A1 US2014032527 A1 US 2014032527A1
Authority
US
United States
Prior art keywords
parallel
operations
deferred
map
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US14/033,145
Other versions
US8959499B2 (en
Inventor
Craig D. Chambers
Ashish Raniwala
Frances J. Perry
Stephen R. Adams
Robert R. Henry
Robert Bradshaw
Nathan Weizenbaum
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Google LLC
Original Assignee
Google LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US14/033,145 priority Critical patent/US8959499B2/en
Application filed by Google LLC filed Critical Google LLC
Assigned to GOOGLE INC. reassignment GOOGLE INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHAMBERS, CRAIG D., HENRY, ROBERT R., ADAMS, STEPHEN R., BRADSHAW, ROBERT, PERRY, FRANCES J., RANIWALA, ASHISH, WEIZENBAUM, NATHAN
Publication of US20140032527A1 publication Critical patent/US20140032527A1/en
Priority to US14/622,556 priority patent/US9626202B2/en
Application granted granted Critical
Publication of US8959499B2 publication Critical patent/US8959499B2/en
Priority to US15/483,044 priority patent/US10133592B2/en
Assigned to GOOGLE LLC reassignment GOOGLE LLC CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: GOOGLE INC.
Priority to US16/175,925 priority patent/US10338942B2/en
Priority to US16/449,987 priority patent/US10795705B2/en
Priority to US17/009,420 priority patent/US11392398B2/en
Priority to US17/834,256 priority patent/US11755351B2/en
Priority to US18/229,450 priority patent/US20230376332A1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45504Abstract machines for programme code execution, e.g. Java virtual machine [JVM], interpreters, emulators
    • G06F17/30445
    • G06F17/30471
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/57Certifying or maintaining trusted computer platforms, e.g. secure boots or power-downs, version controls, system software checks, secure updates or assessing vulnerabilities
    • G06F21/577Assessing vulnerabilities and evaluating computer system security
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F8/00Arrangements for software engineering
    • G06F8/30Creation or generation of source code
    • G06F8/31Programming languages or programming paradigms
    • G06F8/314Parallel programming languages
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F8/00Arrangements for software engineering
    • G06F8/30Creation or generation of source code
    • G06F8/34Graphical or visual programming
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F8/00Arrangements for software engineering
    • G06F8/40Transformation of program code
    • G06F8/41Compilation
    • G06F8/43Checking; Contextual analysis
    • G06F8/433Dependency analysis; Data or control flow analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • G06F9/3836Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution
    • G06F9/3851Instruction issuing, e.g. dynamic instruction scheduling or out of order instruction execution from multiple instruction streams, e.g. multistreaming
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • G06F9/3885Concurrent instruction execution, e.g. pipeline, look ahead using a plurality of independent parallel functional units
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/445Program loading or initiating
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/455Emulation; Interpretation; Software simulation, e.g. virtualisation or emulation of application or operating system execution engines
    • G06F9/45533Hypervisors; Virtual machine monitors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/48Program initiating; Program switching, e.g. by interrupt
    • G06F9/4806Task transfer initiation or dispatching
    • G06F9/4843Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2453Query optimisation
    • G06F16/24532Query optimisation of parallel queries
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/245Query processing
    • G06F16/2453Query optimisation
    • G06F16/24534Query rewriting; Transformation
    • G06F16/24547Optimisations to support specific applications; Extensibility of optimisers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2221/00Indexing scheme relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F2221/03Indexing scheme relating to G06F21/50, monitoring users, programs or devices to maintain the integrity of platforms
    • G06F2221/034Test or assess a computer or a system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/448Execution paradigms, e.g. implementations of programming paradigms
    • G06F9/4494Execution paradigms, e.g. implementations of programming paradigms data driven

Definitions

  • This disclosure relates to parallel processing of data.
  • Large-scale data processing may include parallel processing, which generally involves performing some operation over each element of a large data set.
  • the various operations may be chained together in a data-parallel pipeline to create an efficient mechanism for processing a data set.
  • a system in one aspect, includes one or more processing devices and one or more storage devices.
  • the storage devices store instructions that, when executed by the one or more processing devices, implement an application, an evaluator, an optimizer, and an executor.
  • the application includes a data parallel pipeline.
  • the data parallel pipeline specifies multiple parallel data objects that contain multiple elements and multiple parallel operations that operate on the parallel data objects.
  • the evaluator is configured, based on the data parallel pipeline, to generate a dataflow graph of deferred parallel data objects and deferred parallel operations corresponding to the data parallel pipeline.
  • Deferred parallel data objects can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object.
  • Deferred parallel operations can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • the optimizer is configured to apply one or more graph transformations to the dataflow graph to generate a revised dataflow graph that includes one or more of the deferred parallel data objects and deferred, combined parallel data operations.
  • the executor configured to execute the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects.
  • Materialized parallel data objects for example, can be data structures that include the data or elements of the parallel data object.
  • the deferred, combined parallel data operations may include at least one generalized mapreduce operation.
  • the generalized mapreduce operation may include multiple, parallel map operations and multiple, parallel reduce operations and be translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations.
  • the executor may be configured to translate the combined mapreduce operation to the single mapreduce operation and execute the single mapreduce operation.
  • the executor may be configured to determine whether to execute the single mapreduce operation as a local, sequential operation or a remote, parallel operation.
  • To translate the generalized mapreduce operation to the single mapreduce operation the executor may be configured to generate a map function that includes the multiple map operations and a reducer function that includes the multiple reducer operations.
  • the multiple parallel data objects may be first class objects of a host programming language.
  • the pipeline further may include a single data object that contains a single element and the dataflow graph includes a corresponding deferred single data object. At least one of the multiple parallel operations in the pipeline may operate on the single data object and one of the multiple parallel data objects and the dataflow graph may include a corresponding deferred parallel operation that operates on a deferred single data object and a deferred parallel data object.
  • the executor may be configured to cache one or more results of the execution of the deferred, combined parallel operations for use in a future execution of the data parallel pipeline.
  • a method in another aspect, includes executing an application that includes a data parallel pipeline.
  • the data parallel pipeline specifies multiple parallel data objects that contain multiple elements and multiple parallel operations that operate on the parallel data objects.
  • the method further includes generating, based on the data parallel pipeline, a dataflow graph of deferred parallel data objects and deferred parallel operations corresponding to the data parallel pipeline.
  • Deferred parallel data objects for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object.
  • Deferred parallel operations can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • the method also includes applying one or more graph transformations to the dataflow graph to generate a revised dataflow graph that includes one or more of the deferred parallel data objects and deferred, combined parallel data operations.
  • the method includes executing the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects.
  • Materialized parallel data objects for example, can be data structures that include the data or elements of the parallel data object.
  • the deferred, combined parallel data operations may include at least one generalized mapreduce operation.
  • the generalized mapreduce operation may include multiple, parallel map operations and multiple, parallel reduce operations and be translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations.
  • Executing the generalized mapreduce operation may include translating the combined mapreduce operation to the single mapreduce operation and executing the single mapreduce operation.
  • Executing the single mapreduce operation may include determining whether to execute the single mapreduce operation as a local, sequential operation or a remote, parallel operation.
  • Translating the generalized mapreduce operation to the single mapreduce operation may include generating a map function that includes the multiple map operations and a reducer function that includes the multiple reducer operations.
  • the multiple parallel data objects may be first class objects of a host programming language.
  • the pipeline further may include a single data object that contains a single element and the dataflow graph includes a corresponding deferred single data object. At least one of the multiple parallel operations in the pipeline may operate on the single data object and one of the multiple parallel data objects and the dataflow graph may include a corresponding deferred parallel operation that operates on a deferred single data object and a deferred parallel data object.
  • a system includes one or more processing devices and one or more storage devices.
  • the storage devices store instructions that, when executed by the one or more processing devices, implement an executor.
  • the executor is configured to access a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations.
  • Deferred parallel data objects for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object.
  • Deferred parallel operations can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • the executor is configured to execute the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects.
  • Materialized parallel data objects can be data structures that include the data or elements of the parallel data object.
  • the executor is configured to execute the at least one deferred, combined parallel operation by determining an estimated size of data associated with the at least one deferred, combined parallel operation; determining whether the estimated size exceeds a threshold size; if the estimated size is below a threshold size, execute the at least one deferred, combined parallel operation as a local, sequential operation; and if the estimated size exceeds a threshold size, execute the at least one deferred, combined parallel operation as remote, parallel operation.
  • Implementations of this aspect may include one or more of the following features.
  • the data associated with the at least one deferred, combined parallel operation may include one or more of input data for the at least one deferred, combined parallel operation, intermediary data produced by the at least one deferred, combined parallel operation, or output data produced by the at least one deferred, combined parallel operation.
  • the at least one deferred, combined parallel data operation may be a generalized mapreduce operation.
  • the generalized mapreduce operation may include multiple, parallel map operations and multiple, parallel reduce operations and be translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations.
  • the executor may be configured to translate the combined mapreduce operation to the single mapreduce operation and execute the single mapreduce operation.
  • the executor may be configured to cause the single mapreduce operation to be copied and executed on multiple, different processing modules in a datacenter.
  • the executor may be configured to generate a map function that includes the multiple map operations and a reducer function that includes the multiple reducer operations.
  • the executor may be configured to access annotations in the dataflow graph that reflect an estimate of the size of the data associated with the at least one deferred, combined parallel operation.
  • a method in another aspect, includes accessing a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations.
  • Deferred parallel data objects for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object.
  • Deferred parallel operations for example, can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • the method also includes executing the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects.
  • Materialized parallel data objects can be data structures that include the data or elements of the parallel data object.
  • executing the at least one deferred, combined parallel operation includes determining an estimated size of data associated with the at least one deferred, combined parallel operation; determining whether the estimated size exceeds a threshold size; if the estimated size is below a threshold size, execute the at least one deferred, combined parallel operation as a local, sequential operation; and if the estimated size exceeds a threshold size, execute the at least one deferred, combined parallel operation as remote, parallel operation.
  • Implementations of this aspect may include one or more of the following features.
  • the data associated with the at least one deferred, combined parallel operation may include one or more of input data for the at least one deferred, combined parallel operation, intermediary data produced by the at least one deferred, combined parallel operation, or output data produced by the at least one deferred, combined parallel operation.
  • the at least one deferred, combined parallel data operation may be a generalized mapreduce operation.
  • the generalized mapreduce operation may include multiple, parallel map operations and multiple, parallel reduce operations and be translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations.
  • Executing the generalized mapreduce operation may include translating the combined mapreduce operation to the single mapreduce operation and executing the single mapreduce operation.
  • Executing the single mapreduce operation as a remote, parallel operation may include causing the single mapreduce operation to be copied and executed on multiple, different processing modules in a datacenter.
  • Translating the generalized mapreduce operation to the single mapreduce operation may include generating a map function that includes the multiple map operations and a reducer function that includes the multiple reducer operations.
  • Determining the estimated size may include accessing annotations in the dataflow graph that reflect an estimate of the size of the data associated with the at least one deferred, combined parallel operation.
  • a system in one aspect, includes one or more processing devices and one or more storage devices.
  • the storage devices store instructions that, when executed by the one or more processing devices, implement an executor.
  • the executor is configured to access a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations.
  • Deferred parallel data objects for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object.
  • Deferred parallel operations can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • At least one of the deferred, combined parallel data operation is a generalized mapreduce operation.
  • the generalized mapreduce operation includes multiple, parallel map operations and multiple, parallel reduce operations and is translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations.
  • the executor is further configured to execute the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects.
  • Materialized parallel data objects for example, can be data structures that include the data or elements of the parallel data object.
  • the executor is configured to translate the combined mapreduce operation to the single mapreduce operation and execute the single mapreduce operation.
  • Implementations of this aspect may include one or more of the following features.
  • the executor may be configured to generate a map function that includes the multiple map operations and a reduce function that includes the multiple reduce operations.
  • the executor may be configured to execute the single mapreduce operation as a remote, parallel operation.
  • the executor may be configured to cause the single mapreduce operation to be executed by multiple worker processes on multiple, different processing modules.
  • the executor may be configured to cause multiple map worker processes to be invoked for each of the map operations, where each of the multiple map worker processes is assigned an index number.
  • Each of the map worker processes may be configured to receive the map function that implements the multiple map operations, one or more inputs associated with one of the map operations, and the map worker process' associated index, select the map operation associated with the inputs based on the map worker process' associated index, and invoke the selected map operation on the one or more inputs.
  • the executor may be configured to cause multiple reduce worker processes to be invoked for each of the reduce operations, where each of the multiple reduce worker processes is assigned an index number.
  • Each of the multiple reduce worker processes may be configured to receive the reduce function that implements the multiple reduce operations, one or more inputs associated with one of the reduce operations, and the reduce worker process' associated index, select the reduce operation associated with the inputs based on the worker process' associated index, and invoke the selected reduce operation on the one or more inputs.
  • a method in another aspect, includes accessing a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations.
  • Deferred parallel data objects for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object.
  • Deferred parallel operations for example, can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet .
  • At least one of the deferred, combined parallel data operation is a generalized mapreduce operation.
  • the generalized mapreduce operation includes multiple, parallel map operations and multiple, parallel reduce operations and being translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations;
  • the method also includes executing the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects.
  • Materialized parallel data objects for example, can be data structures that include the data or elements of the parallel data object.
  • Executing the generalized mapreduce operation includes translating the combined mapreduce operation to the single mapreduce operation and executing the single mapreduce operation.
  • Translating the generalized mapreduce operation to the single mapreduce operation may include generating a map function that includes the multiple map operations and a reduce function that includes the multiple reduce operations.
  • Executing the single mapreduce operation may include executing the single mapreduce operation as a remote, parallel operation.
  • Executing the single mapreduce operation as a remote, parallel operation may include causing the single mapreduce operation to be executed by multiple worker processes on multiple, different processing modules.
  • Causing the single map reduce operation to be executed by multiple worker processes may include causing the multiple map worker processes to be invoked for each of the map operations, where each of the multiple map worker processes is assigned an index number.
  • Each of the map worker processes may be configured to receive the map function that implements the multiple map operations, one or more inputs associated with one of the map operations, and the map worker process' associated index, select the map operation associated with the inputs based on the map worker process' associated index, and invoke the selected map operation on the one or more inputs.
  • Causing the single map reduce operation to be executed by multiple worker processes may include causing multiple reduce worker processes to be invoked for each of the reduce operations, where each of the multiple reduce worker processes is assigned an index number.
  • Each of the multiple reduce worker processes may be configured to receive the reduce function that implements the multiple reduce operations, one or more inputs associated with one of the reduce operations, and the reduce worker process' associated index, select the reduce operation associated with the inputs based on the worker process' associated index, and invoke the selected reduce operation on the one or more inputs.
  • FIG. 1 is a block diagram illustrating an example of a datacenter.
  • FIG. 2 is a block diagram of an example of a processing module.
  • FIG. 3 is a block diagram illustrating an example of a pipeline library.
  • FIG. 4A is a flow chart illustrating an example of a process that may be performed by an evaluator, an optimizer, and an executor of the pipeline library.
  • FIG. 4B is a flow chart illustrating an example of a process that may be performed by the executor of the pipeline library.
  • FIGS. 5A and 5B show an example dataflow graph transformation that illustrates ParallelDo producer-consumer fusion and sibling fusion.
  • FIGS. 6A and 6B show an example dataflow graph transformation that illustrates MSCR fusion.
  • FIGS. 7A-7E illustrate an example of a dataflow graph transformation performed to generate a final dataflow graph.
  • FIG. 8 illustrates an example of an MSCR operation with 3 input channels.
  • the techniques described in this document can be applied to large-scale data processing and, in particular, to large scale data-parallel pipelines.
  • large-scale processing may be performed in a distributed data processing system, such as a datacenter or a network of datacenters.
  • a distributed data processing system such as a datacenter or a network of datacenters.
  • large-scale Internet services and the massively parallel computing infrastructure that support such services may employ warehouse-sized computing systems, made up of thousands or tens of thousands of computing nodes.
  • FIG. 1 is a block diagram illustrating an example of a datacenter 100 .
  • the datacenter 100 is used to store data, perform computational tasks, and transmit data to other systems outside of the datacenter using, for example, a network connected to the datacenter.
  • the datacenter 100 may perform large-scale data processing on massive amounts of data.
  • the datacenter 100 includes multiple racks 102 . While only two racks are shown, the datacenter 100 may have many more racks.
  • Each rack 102 can include a frame or cabinet into which components, such as processing modules 104 , are mounted.
  • each processing module 104 can include a circuit board, such as a motherboard, on which a variety of computer-related components are mounted to perform data processing.
  • the processing modules 104 within each rack 102 are interconnected to one another through, for example, a rack switch, and the racks 102 within each datacenter 100 are also interconnected through, for example, a datacenter switch.
  • the processing modules 104 may each take on a role as a master or slave.
  • the master modules control scheduling and data distribution tasks amongst themselves and the slaves.
  • a rack can include storage (e.g., one or more network attached disks) that is shared by the one or more processing modules 104 and/or each processing module 104 may include its own storage. Additionally, or alternatively, there may be remote storage connected to the racks through a network.
  • the datacenter 100 may include dedicated optical links or other dedicated communication channels, as well as supporting hardware, such as modems, bridges, routers, switches, wireless antennas and towers, and the like.
  • the datacenter 100 may include one or more wide area networks (WANs) as well as multiple local area networks (LANs).
  • WANs wide area networks
  • LANs local area networks
  • FIG. 2 is a block diagram of an example of a processing module 200 , which may be used for one or more of the processing modules 104 .
  • the processing module 200 includes memory 202 , one or more processing units (CPUs) 204 , and one or more network or other communication interfaces 206 . These components are interconnected by one or more communication buses.
  • the processing module 200 may include an input/output (I/O) interface connecting the processing module to input and output devices such as a display and a keyboard.
  • Memory 202 may include high speed random access memory and may also include non-volatile memory, such as one or more magnetic disk storage devices. Memory 202 may include mass storage that is remotely located from the CPU 204 .
  • the memory 202 stores application software 202 a, a mapreduce library 202 b , a pipeline library 202 c , and an operating system 202 d (e.g., Linux).
  • the operating system 202 d generally includes procedures for handling various basic system services and for performing hardware dependent tasks.
  • the application software 202 a performs large-scale data processing.
  • the libraries 202 b and 202 c provide functions and classes that may be employed by the application software 202 a to perform large-scale data processing and implement data-parallel pipelines in such large-scale data processing.
  • the mapreduce library 202 b can support the MapReduce programming model for processing massive amounts of data in parallel.
  • the MapReduce model is described in, for example, MapReduce: Simplified Data Processing on Large Clusters, OSDI'04: Sixth Symposium on Operating System Design and Implementation, San Francisco, Calif., December, 2004 and U.S. Pat. No. 7,650,331, both of which are incorporated by reference.
  • the MapReduce model provides an abstraction to application developers for how to think about their computations.
  • the application developers can formulate their computations according to the abstraction, which can simplify the building of programs to perform large-scale parallel-data processing.
  • the application developers can employ the MapReduce model with or without using the mapreduce library 202 b.
  • the mapreduce library 202 b can manage many of the difficult low-level tasks. Such low-level tasks may include, for example, selecting appropriate parallel worker machines, distributing to them the program to run, managing the temporary storage and flow of intermediate data between the three phases, synchronizing the overall sequencing of the phases, and coping with transient failures of machines, networks, and software.
  • the MapReduce model generally involves breaking computations down into a mapreduce operation, which includes a single map operation and a single reduce operation.
  • the map operation performs an operation on each of the logical records in the input to compute a set of intermediate key/value pairs.
  • the reduce operation performs an operation on the values that share the same key to combine the values in some manner. Implicit in this model is a shuffle operation, which involves grouping all of the values with the same key.
  • the mapreduce library 202 b may implement a map phase, a shuffle phase, and a reduce phase to support computations formulated according to the MapReduce model.
  • a user program or another library, such as pipeline library 202 c ) calls the mapreduce library 202 b , specifying information identifying the input file(s), information identifying or specifying the output files to receive output data, and two application-specific data processing operators, the map operator and the reduce operator.
  • the map operator specifies a map function that processes the input data to produce intermediate data
  • the reduce operator specifies a reduce function that merges or otherwise combines the intermediate data values.
  • the mapreduce library 202 b then employs this information to implement that map phase, the shuffle phase, and the reduce phase.
  • the map phase starts by reading a collection of values or key/value pairs from an input source, such as a text file, binary record-oriented file, or MySq 1 database. Large data sets may be represented by multiple, even thousands, of files (which may be referred to as shards), and multiple file shards can be read as a single logical input source.
  • the map phase then invokes the user-defined function, the map function or Mapper, on each element, independently and in parallel. For each input element, the user-defined function emits zero or more key/value pairs, which are the outputs of the map phase.
  • the shuffle phase takes the key/value pairs emitted by the Mappers and groups together all the key/value pairs with the same key. The shuffle phase then outputs each distinct key and a stream of all the values with that key to the next phase, the reduce phase.
  • the reduce phase takes the key-grouped data emitted by the shuffle phase and invokes the user-defined function, the reduce function or Reducer, on each distinct key-and-values group, independently and in parallel.
  • Each Reducer invocation is passed a key and an iterator over all the values associated with that key, and emits zero or more replacement values to associate with the input key.
  • the Reducer typically performs some kind of aggregation over all the values with a given key. For some operations, the Reducer is just the identity function.
  • the key/value pairs emitted from all the Reducer calls are then written to an output sink, e.g., a sharded file or database.
  • the mapreduce library 202 b may divide the input pieces into M pieces (for example, into 64 megabyte (MB) sized files) and start up multiple copies of the program that uses the library 202 b on a cluster of machines, such as multiple ones of the processing modules 104 .
  • One of the copies may be a master copy and the rest may be worker copies that are assigned work by the master.
  • the master selects idle workers and assigns each one a map task or a reduce task.
  • the workers assigned to a map task use the Mapper to perform the mapping operation on the inputs to produce the intermediate results, which are divided, for example, into R sets. When the intermediate results are divided into R sets, there are R reduce tasks to assign.
  • the workers assigned to a reduce task use the Reducer to perform the reduce operation on the intermediate values to produce the output.
  • the master returns to the user program or library employing the mapreduce library 202 b.
  • the mapreduce operation is implemented as a set of parallel operations across a cluster of processing devices.
  • a separate user-defined Combiner function can be specified to perform partial combining of values associated with a given key during the map phase.
  • Each map worker can keep a cache of key/value pairs that have been emitted from the Mapper, and use the Combiner function to combine locally as much as possible before sending the combined key/value pairs on to the Shuffle phase.
  • the Reducer may complete the combining step by combining values from different map workers.
  • the Shuffle phase may send each key-and-values group to arbitrarily but deterministically chosen reduce worker machine, with this choice determining which output file shard will hold that key's results.
  • a user defined Sharder function can be specified that selects which reduce worker machine should receive the group for a given key.
  • a user-defined Sharder can be used to aid in load balancing.
  • the user-defined Sharder can also be used to sort the output keys into reduce “buckets,” with all the keys of the i th reduce worker being ordered before all the keys of the i th +1st reduce worker. Coupled with the fact that each reduce worker processes keys in lexicographic order, this kind of Sharder can be used to produce sorted output.
  • the pipeline library 202 c provides functions and classes that support data-parallel pipelines and, in particular, pipelines that include chains or directed acyclic graphs of mapreduce operations.
  • the pipeline library 202 c may help alleviate some of the burdens of implementing chains of mapreduce operations.
  • many real-world computations require a chain of mapreduce stages. While some logical computations can be expressed as a mapreduce operation, others require a sequence or graph of mapreduce operations. As the complexity of the logical computations grows, the challenge of mapping the computations into physical sequences of mapreduce operations increases. Higher-level concepts such as “count the number of occurrences” or “join tables by key” are generally hand-compiled into lower-level mapreduce operations.
  • the user may take on the additional burdens of writing a driver program to invoke the mapreduce operations in the proper sequence, and managing the creation and deletion of intermediate files holding the data.
  • the pipeline library 202 c may obviate or reduce some of the difficulty in producing data-parallel pipelines that involve multiple mapreduce operations, as well as the need for the developer to produce additional coordination code to chain together the separate mapreduce stages in such data-parallel pipelines.
  • the pipeline library 202 c also may obviate or reduce additional work to manage the creation and later deletion of the intermediate results in between pipeline stages.
  • the pipeline library 202 c may help prevent the logical computation itself from becoming hidden among all the low-level coordination details, thereby making it easier for new developers to understand the computation.
  • making use of the pipeline library 202 c may help prevent the division of the pipeline into particular stages from becoming “baked in” to the code and difficult to change later if the logical computation needs to evolve.
  • the application software 202 a may employ one or both of the libraries 202 b or 202 c .
  • An application developer may develop application software that employs the mapreduce library 202 b to perform computations formulated as a mapreduce operation.
  • the application developer may alternatively, or additionally, employ the pipeline library 202 c when developing a data-parallel pipeline that includes multiple mapreduce operations.
  • the pipeline library 202 c may allow the developer to code the computations in a more natural manner, using the native programming language in which the pipeline library 202 c is implemented, without thinking about casting the logical computation in terms of mapreduce operations or building an ordered graph of operations.
  • the pipeline library 202 c can formulate the logical computation in terms of multiple mapreduce operations prior to execution, and then execute the computation either by implementing the mapreduce operations itself, or interfacing with the mapreduce library 202 b to implement the mapreduce operations.
  • FIG. 3 is a block diagram illustrating an example of a pipeline library 300 that may be used to implement pipeline library 200 c.
  • the pipeline library 300 includes one or more parallel data collection classes 302 , one or more parallel operations 304 , an evaluator 306 , an optimizer 308 , and an executor 310 .
  • the parallel data collection classes 302 are used to instantiate parallel data objects that hold a collection of data
  • the parallel operations 304 are used to perform parallel operations on the data held by the parallel data objects.
  • the parallel operations 304 may be composed to implement data-parallel computations and an entire pipeline, or even multiple pipelines, can be implemented using the parallel collection classes 302 and parallel operations 304 .
  • Parallel data collection classes 302 and operations 304 present a simple, high-level, uniform abstraction over many different data representations and over different execution strategies.
  • the parallel data collection classes 302 abstract away the details of how data is represented, including whether the data is represented as an in-memory data structure, as one or more files, or as an external storage service.
  • parallel operations 304 abstract away their implementation strategy, such as whether an operation is implemented as a local, sequential loop, as a remote parallel invocation of the mapreduce library 202 b , as a query on a database, or as a streaming computation.
  • the evaluator 306 defers the evaluation of parallel operations. Instead, the evaluator 306 constructs an internal execution plan dataflow graph that contains the operations and their arguments. Once the execution plan dataflow graph for the whole logical computation is constructed, the optimizer 308 revises the execution plan, for example, by applying graph transformations that fuse or combine chains of parallel operations together into a smaller number of combined operations.
  • the revised execution plan may include a generalized mapreduce operation that includes multiple, parallel map operations and multiple, parallel reduce operations (for example, the MapShuffleCombineReduce operation described further below), but which can be translated to a single mapreduce operation with a single map function to implement the multiple map operations and a single reduce function to implement the multiple reduce operations.
  • the executor 310 executes the revised operations using underlying primitives (e.g., MapReduce operations). When running the execution plan, the executor 310 may choose which strategy to use to implement each operation (e.g., local sequential loop vs. remote parallel MapReduce) based in part on the size of the data being processed.
  • the executor 310 also may place remote computations near the data on which they operate, and may perform independent operations in parallel.
  • the executor 310 also may manage the creation and cleanup of any intermediate files needed within the computation.
  • the pipeline library 300 may be implemented in any of a number of programming languages. The following describes examples of aspects of an implementation in the Java programming language.
  • the pipeline library 300 provides a parallel data collection class referred to as a PCollection ⁇ T>, which is an immutable bag of elements of type T.
  • a PCollection can either have a well-defined order (called a sequence), or the elements can be unordered (called a collection). Because they are less constrained, collections may be more efficient to generate and process than sequences.
  • a PCollection ⁇ T> can be created by reading a file in one of several possible formats. For example, a text file can be read as a PCollection ⁇ String>, and a binary record-oriented file can be read as a PCollection ⁇ T>, given a specification of how to decode each binary record into an object of type T.
  • a PCollection ⁇ T> may also be created from an in-memory Java Collection ⁇ T>.
  • Data sets represented by multiple file shards can be read in as a single logical PCollection. For example:
  • recordsOf( . . . ) specifies a particular way in which a DocInfo instance is encoded as a binary record.
  • Other predefined encoding specifiers may include strings( ) for UTF-8-encoded text, ints( ) for a variable-length encoding of 32-bit integers, and pairsOf(e 1 ,e 2 ) for an encoding of pairs derived from the encodings of the components.
  • Some implementations may allow users to specify their own custom encodings.
  • a second parallel data collection class 302 is PTable ⁇ K,V>, which represents an (immutable multi-map with keys of type K and values of type V.
  • PTable ⁇ K,V> may be just an unordered bag of pairs.
  • Some of the parallel operations 304 may apply only to PCollections of pairs, and in Java PTable ⁇ K,V> may be implemented as a subclass of PCollection ⁇ Pair ⁇ K,V>> to capture this abstraction.
  • PTable ⁇ K,V> might be defined as a type synonym of PCollection ⁇ Pair ⁇ K,V>>.
  • the parallel data objects may be implemented as first class objects of the native language in which the library 300 is implemented.
  • the objects may be manipulable like other objects in the native language.
  • the PCollections may be able to be passed into and returned from regular methods in the language, and may be able to be stored in other data structures of the language (although some implementations may prevent the PCollections from being stored in other PCollections).
  • regular control flow constructs of the native language may be able to be used to define computations involving objects, including functions, conditionals, and loops. For example, if Java is the native language:
  • Implementing the parallel data objects as first class objects in the native language of the library 300 may simplify the development of programs using the library, since the developer can use the parallel data objects in the same manner he or she would use other objects.
  • the pipeline library 300 can also include a single data collection class PObject ⁇ T> to support the ability to inspect the contents of PCollections during the execution of a pipeline.
  • a PObject ⁇ T> is a container for a single object of type T (for example, a single native object (e.g., Java object) of type T) and any associated methods of PObjects are designed to operate on a single element.
  • PObjects can be either deferred or materialized (as described further below), allowing them to be computed as results of deferred operations in pipelines. Once a pipeline is run, the contents of a now-materialized PObject can be extracted using getValue( ).
  • an asSequentialCollection( ) operation can be applied to a PCollection ⁇ T> to yield a PObject ⁇ Collection ⁇ T>>, which can be inspected once the pipeline runs to read out all the elements of the computed PCollection as a regular Java in-memory Collection:
  • the combine( ) operation (described below) applied to a PCollection ⁇ T> and a combining function over Ts yields a PObject ⁇ T> representing the fully combined result.
  • Global sums and maxima can be computed this way.
  • the contents of PObjects also may be able to be examined within the execution of a pipeline, for example, using an operate( ) primitive provided by the pipeline library 300 .
  • the operate( ) primitive takes a list of PObjects and an argument OperateFn (which defines the operation to be performed on each PObject), and returns a list of PObjects.
  • OperateFn which defines the operation to be performed on each PObject
  • the OperateFn returns a list of native objects, such as Java objects, and operate( ) wraps these native objects inside of PObjects, which are returned as the results.
  • arbitrary computations can be embedded within a pipeline and executed in deferred fashion.
  • operations other than ParallelDo operations (described below), which operate on PCollections that contain multiple elements, can be included in the pipeline. For example, consider embedding a call to an external service that reads
  • This example uses operations for converting between PCollections and PObjects containing file names.
  • the viewAsFile( ) operation applied to a PCollection and a file format choice yields a PObject ⁇ String> containing the name of a temporary sharded file of the chosen format where the PCollection's contents may be found during execution of the pipeline.
  • File-reading operations such as readRecordFileCollection( ) may be overloaded to allow reading files whose names are contained in PObjects.
  • a DoFn performs an operation on each element of a PCollection, and just receives the PCollection as an input.
  • the operation on each PCollection may involve a value or other data stored in a PObject.
  • the DoFn may receive the PCollection as an input, as normal, and a PObject as a side input.
  • data-parallel operations 304 are invoked on parallel data objects, such as PCollections.
  • the pipeline library 300 defines some primitive data-parallel operations, with other operations being implemented in terms of these primitives.
  • One of the data-parallel primitives is parallelDo( ), which supports elementwise computation over an input PCollection ⁇ T> to produce a new output PCollection ⁇ S>.
  • This operation takes as its main argument a DoFn ⁇ T, S>, a function-like object defining how to map each value in the input PCollection ⁇ T> into zero or more values to appear in the output PCollection ⁇ S>.
  • This operation also takes an indication of the kind of PCollection or PTable to produce as a result. For example:
  • collectionOf(strings( )) specifies that the parallelDo( ) operation should produce an unordered PCollection whose String elements should be encoded using UTF-8.
  • Other options may include sequenceOf(elemEncoding) for ordered PCollections and tableOf(keyEncoding, valueEncoding) for PTables.
  • emitFn is a call-back function passed to the user's process( . . . ) method, which should invoke emitFn.emit(outElem) for each outElem that should be added to the output PCollection.
  • Subclasses of DoFn may be included, such as MapFn (implementing a map) and FilterFn (implementing a filter) to provide simpler interfaces in some cases.
  • the operation parallelDo( ) can be used to express both the map and reduce parts of a MapReduce operation.
  • the library 300 also may include a version of parallelDo( ) that allows multiple output PCollections to be produced simultaneously from a single traversal of the input PCollection.
  • DoFn functions may be prevented from accessing any global mutable state of the enclosing program if DoFn functions can be distributed remotely and run in parallel.
  • DoFn objects may be able to maintain local instance variable state, but there may be multiple DoFn replicas operating concurrently with no shared state.
  • a second primitive, groupByKey( ), converts a multimap of type PTable ⁇ K,V> (which can have many key/value pairs with the same key) into a uni-map of type PTable ⁇ K, Collection ⁇ V>> where each key maps to an unordered collection of all the values with that key. For example, the following computes a table mapping URLs to the collection of documents that link to them:
  • the operation groupByKey( ) corresponds to the shuffle step of MapReduce. There may also be a variant that allows specifying a sorting order for the collection of values for each key.
  • a third primitive, combineValues( ), takes an input PTable ⁇ K, Collection ⁇ V>> and an associative combining function on Vs, and returns a PTable ⁇ K,V> where each input collection of values has been combined into a single output value. For example:
  • the operation combineValues( ) is semantically a special case of parallelDo( ), but the associativity of the combining function allows the operation to be implemented through a combination of a MapReduce Combiner (which runs as part of each mapper) and a MapReduce Reducer (to finish the combining), which may be more efficient than doing all the combining in the reducer.
  • a fourth primitive, flatten( ) takes a list of PCollection ⁇ T>s and returns a single PCollection ⁇ T> that contains all the elements of the input PCollections.
  • the operation flatten( ) may not actually copy the inputs, but rather just view the inputs as if the inputs were one logical PCollection.
  • a pipeline typically concludes with operations that write the final resulting PCollections to external storage. For example:
  • the pipeline library 300 may include a number of other operations on PCollections that are derived in terms of the above-described primitives. These derived operations may be the same as helper functions the user could write. For example, a count( ) operation takes a PCollection ⁇ T> and returns a PTable ⁇ T,Integer> mapping each distinct element of the input PCollection to the number of times the element occurs. This function may be implemented in terms of parallelDo( ), groupByKey( ), and combineValues( ), using the same pattern as was used to compute wordCounts above.
  • the code above can be simplified to the following:
  • join( ) implements a join over two or more PTables sharing a common key type.
  • join( ) returns a unimap PTable ⁇ K, Pair ⁇ Collection ⁇ V 1 >, Collection ⁇ V 2 >>> that maps each key in either of the input tables to the collection of all values with that key in the first table, and the collection of all values with that key in the second table.
  • This resulting table can be processed further to compute a traditional inner or outer-join, but it may be more efficient to be able to manipulate the value collections directly without computing their cross-product.
  • the operation join( ) may be implemented as follows:
  • top( ) Another derived operation is top( ) which takes a comparison function and a count N and returns the greatest N elements of its receiver PCollection according to the comparison function.
  • This operation may be implemented on top of parallelDo( ), groupByKey( ), and combineValues( ).
  • the pipeline library 300 executes parallel operations lazily, using deferred evaluation.
  • the evaluator 306 defers the evaluation of parallel operations, and instead constructs an internal execution plan dataflow graph that contains the operations and the arguments of the operations.
  • Each parallel data object such as a PCollection, is represented internally either in deferred (not yet computed) or materialized (computed) state.
  • a deferred parallel data object for example, holds a pointer to the deferred operation that computes the parallel data object.
  • a deferred operation may hold references to the parallel data objects that are the arguments of the deferred operation (which may themselves be deferred or materialized) and the deferred parallel data objects that are the results of the operation.
  • the library 300 When a library operation like parallelDo( ) is called, the library 300 creates a ParallelDo deferred operation object and returns a new deferred PCollection that points to the operation.
  • the evaluator 306 converts the parallel data objects and parallel operations into a directed acyclic graph of deferred (unevaluated) objects and operations. This graph may be referred to as the execution plan or execution plan dataflow graph.
  • the optimizer 308 fuses chains or subgraphs of parallel operations in the dataflow graph together into a smaller number of operations (some of which may be combined operations), which the executor 310 can then execute using an underlying primitive or other logic.
  • the optimizer 308 may be written, for example, as a series of independent graph transformations. In one implementation, the optimizer 308 performs a series of passes over the initial execution plan that reduces the number of overall operations and groups operations, with the overall goal of producing the fewest MapShuffleCombineReduce (MSCR) operations.
  • MSCR MapShuffleCombineReduce
  • An MSCR operation includes a combination of ParallelDo, GroupByKey, CombineValues, and Flatten operations.
  • An MSCR operation can be mapped to and run as a single mapreduce operation.
  • An MSCR operation has M input channels (each performing a map operation) and R output channels (each performing a shuffle, a combine, and a reduce).
  • Each input channel m takes a PCollection ⁇ T m > as input and performs an R-output ParallelDo “map” operation on that input to produce R outputs of type PTable ⁇ K r ,V r >s.
  • Each output channel R flattens its M inputs and then either (a) performs a GroupByKey “shuffle,” an optional CombineValues “combine,” and a O r -output ParallelDo “reduce” (which defaults to the identity operation), and then writes the results to O r output PCollections or (b) writes the input directly as the output.
  • the former kind of output channel may be referred to as a “grouping” channel, while the latter kind of output channel may be referred to as a “pass-through” channel.
  • a pass-through channel may allow the output of a mapper be a result of an MSCR operation.
  • FIG. 8 illustrates an example of an MSCR operation 800 with 3 input channels 802 a, 802 b, and 802 c .
  • the first input channel 802 a performs a ParallelDo Ml 804 a .
  • the second input channel 802 b performs a ParallelDo M 2 804 b .
  • the third input channel 802 c performs a ParallelDo M 3 804 c .
  • the MSCR operation includes two grouping output channels 806 a and 806 b .
  • the first grouping output channel 806 a includes a GroupByKey GBK 1 808 a, CombineValues CV 1 810 a , and a reducing ParallelDo R 1 812 a.
  • the second grouping output channel includes a GroupByKey GBK 2 808 b , CombineValues CV 2 810 b, and a reducing ParallelDo R 2 812 b .
  • the MSCR operation 800 also includes one pass-through output channel 814 .
  • MSCR generalizes the MapReduce model by allowing multiple mappers and multiple reducers and combiners, by allowing each reducer to produce multiple outputs, by removing the requirement that the reducer must produce outputs with the same key as the reducer input, and by allowing pass-through outputs.
  • any given MSCR may include multiple, parallel map operations that each operate on different inputs and multiple reduce operations that operate on the outputs of the map operations to produce multiple different outputs.
  • each MSCR operation can be implemented using a single mapreduce operation that includes a single map function to implement the map operations on the different inputs and a single reduce function to implement the reduce operations to produce the multiple outputs.
  • the executor 310 executes the revised execution plan dataflow graph.
  • the pipeline library 300 performs batch execution.
  • the executor 310 traverses the operations in the revised execution plan in forward topological order, and executes each one in turn. Independent operations may be able to be executed simultaneously.
  • incremental or continuous execution of pipelines may be implemented, where incrementally added inputs lead to quick, incremental update of outputs. Further, optimization may be performed across pipelines run by multiple users over common data sources.
  • the executor 310 executes operations other than a MSCR by performing the appropriate computations that perform the operation. MSCRs are mapped to a single mapreduce operation, which is then executed.
  • the executor 310 first decides whether the mapreduce operation should be run locally and sequentially, or as a remote, parallel mapreduce operation (using, for example, mapreduce library 202 b ). Since there is overhead in launching a remote, parallel job, local evaluation may be used for modest-size inputs where the gain from parallel processing is outweighed by the start-up overheads. Modest-size data sets may be common during development and testing. Using local evaluation for these data sets may therefore facilitate the use of regular IDEs, debuggers, profilers, and related tools, easing the task of developing programs that include data-parallel computations.
  • the executor 310 may choose to launch a remote, parallel MapReduce operation using the mapreduce library 202 b .
  • the executor 310 may use observations of the input data sizes and estimates of the output data sizes to automatically choose a reasonable number of parallel worker machines. Users can assist in estimating output data sizes, for example by augmenting a DoFn with a method that returns the expected ratio of output data size to input data size, based on the computation represented by that DoFn. Estimates may be refined through dynamic monitoring and feedback of observed output data sizes. Relatively more parallel workers may be allocated to jobs that have a higher ratio of CPU to I/O.
  • the executor 310 may automatically create temporary files to hold the outputs of each operation executed. Once the pipeline is completed, all of these temporary files may be automatically deleted. Alternatively, or additionally, some or all of these temporary files may be deleted as soon as they are no longer needed later in the pipeline.
  • the pipeline library 300 may be designed to make building and running pipelines feel as similar as possible to running a regular program in the native language for which the pipeline library was designed.
  • the native language is Java
  • using local, sequential evaluation for modest-sized inputs is one way to do so.
  • Another way is by automatically routing any output to System.out or System.err from within a user's DoFn, such as debugging prints, from the corresponding remote MapReduce worker to the main program's output streams.
  • DoFn System.err from within a user's DoFn, such as debugging prints
  • any exceptions thrown within a DoFn running on a remote MapReduce worker are captured, sent to the main program, and rethrown.
  • the library 300 may support a cached execution mode. In this mode, rather than recompute an operation, the executor 310 first attempts to reuse the result of that operation from the previous run, if it was saved in a (internal or user-visible) file and if the executor 310 determines that the operation's result hasn't changed. An operation's result may be considered unchanged if (a) the operation's inputs haven't changed, and (b) the operation's code and captured state haven't changed.
  • the executor 310 may perform an automatic, conservative analysis to identify when reuse of previous results is guaranteed to be safe. Caching can lead to quick edit-compile-run-debug cycles, even for pipelines that would normally take hours to run. This may reduce the amount of time required to find a bug in a late pipeline stage, fix the program, and then reexecute the revised pipeline from scratch.
  • FIG. 4A is a flow chart illustrating an example of a process 400 that may be performed by the evaluator 306 , the optimizer 308 , and the executor 310 .
  • the evaluator 306 Based on a data parallel pipeline that includes multiple parallel data objects and multiple parallel data operations that operate on the objects, the evaluator 306 generates a dataflow graph of deferred parallel data objects and deferred parallel operations corresponding to the data parallel pipeline ( 402 ).
  • a deferred parallel data object is one that has not yet been computed and a deferred parallel operation is one that has not been executed.
  • the evaluator 306 may generate a data structure that holds a pointer to the parallel data operation that operates on the parallel data object.
  • the evaluator 306 may generate a data structure that holds a pointer to a parallel data object that is an input to the deferred parallel operation and a pointer to a deferred parallel object that is an output of the deferred parallel operation.
  • the optimizer 308 applies one or more graph transformations to the dataflow graph to generate a revised dataflow graph that includes the deferred parallel data objects (or a subset) and the deferred, combined parallel data operations ( 404 ).
  • the deferred, combined parallel data operations may include one or more generalized mapreduce operations (for example, an MSCR), which includes multiple map operations and multiple reduce operations, but is translatable to a single mapreduce operation that includes a single map function to implement the map operations and a single reduce function to implement the reduce operations.
  • the optimizer 308 performs a series of passes over the dataflow graph, applying the following graph transformations or annotations in the following order: (1) sink flattens; (2) lift CombineValues operations; (3) insert fusion blocks; (4) fuse ParallelDos; and (5) fuse MSCRs.
  • the sink flattens transformation involves pushing a Flatten operation down through consuming ParallelDo operations by duplicating the ParallelDo before each input to the flatten.
  • h(f(a)+g(b)) is equivalent to h(f(a))+h(g(b)).
  • This transformation creates opportunities for ParallelDo fusion (described below).
  • the lift CombineValues operations annotation involves marking certain CombineValues operations for treatment as ParallelDos for ParallelDo fusion. If a CombineValues operation immediately follows a GroupByKey operation, the GroupByKey records that fact. The original CombineValues is left in place, and is henceforth treated as a normal ParallelDo operation and subject to ParallelDo fusion.
  • the insert fusion blocks annotation involves annotating the ParallelDos connecting two GroupByKey operations. If two GroupByKey operations are connected by a chain of one or more ParallelDo operations, the optimizer 308 chooses which ParallelDos should fuse up into the output channel of the earlier GroupByKey, and which should fuse down into the input channel of the later GroupByKey. The optimizer estimates the size of the intermediate PCollections along the chain of ParallelDos, identifies one with minimal expected size, and marks that intermediate PCollection as a boundary blocking ParallelDo fusion (that is, marks the ParallelDos on either side of that PCollection as not being subject to fusion into one another).
  • the fuse ParallelDos transformation involves fusing ParallelDos together.
  • One type of ParallelDo fusion that the optimizer 306 may perform is referred to as producer-consumer fusion. If one ParallelDo operation performs function f, and the result is consumed by another ParallelDo operation that performs function g, the two ParallelDo operations may be replaced by a single ParallelDo that computes both f and g f If the result of the f ParallelDo is not needed by other operations in the graph, fusion has rendered it unnecessary, and the code to produce it may be removed as dead.
  • ParallelDo sibling fusion may be applied when two or more ParallelDo operations read the same input PCollection.
  • the ParallelDo operations can be fused into a single multi-output ParallelDo operation that computes the results of all the fused operations in a single pass over the input.
  • Both producer-consumer and sibling fusion can apply to arbitrary trees of multi-output ParallelDo operations.
  • CombineValues operations are special cases of ParallelDo operations that can be repeatedly applied to partially computed results.
  • ParallelDo fusion may also be applied to CombineValues operations.
  • the fuse MSCRs transformation involves creating MSCR operations.
  • An MSCR operation starts from a set of related GroupByKey operations.
  • GroupByKey operations may be considered related if the operations consume (possibly via Flatten operations) the same input or inputs created by the same ParallelDo operations.
  • the MSCR's input and output channels are derived from the related GroupByKey operations and the adjacent operations in the execution plan.
  • Each ParallelDo operation with at least one output consumed by one of the GroupByKey operations (possibly via Flatten operations) is fused into the MSCR, forming a new input channel. Any other inputs to the GroupByKeys also form new input channels with identity mappers.
  • Each of the related GroupByKey operations starts an output channel.
  • any remaining ParallelDo operations are also transformed into trivial MSCR operations with a single input channel containing the ParallelDo and an identity output channel.
  • the final optimized execution plan contains only MSCR, Flatten, and Operate operations.
  • the executor 310 executes the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects ( 406 ).
  • Executing the generalized mapreduce operation can include translating the generalized mapreduce operation to the single mapreduce operation and executing the single mapreduce operation.
  • the executor 310 may decide whether to execute the single mapreduce operation as a local, sequential operation or a remote, parallel operation and then execute the single mapreduce accordingly. For example, the executor 310 may decide based on the size of the input data set, as described above.
  • FIG. 4B is a flow chart illustrating an example of a process 450 that may be performed by the executor 310 of the pipeline library 202 c to execute the revised dataflow graph.
  • the executor 310 accesses the revised data flow graph ( 452 ) and begins traversing the data flow graph, for example, in a forward topological manner ( 454 ).
  • the executor 310 may support incremental or continuous execution of pipelines.
  • the executor 310 executes those operations locally using logic included in the pipeline library 202 c ( 458 ).
  • the executor 310 determines whether to execute the MSCR as a local, sequential operation or, instead, as a remote, parallel operation using the mapreduce library 202 b ( 460 ). For example, the executor 310 may determine an estimated size of data associated with the MSCR and determine whether the estimated size exceeds a threshold size. If the estimated size is below the threshold size, executor 310 may execute the MSCR as a local, sequential operation ( 462 ).
  • the executor 310 may execute the MSCR operation as remote, parallel operation by translating the MSCR into a single mapreduce operation and executing that mapreduce operation as a remote, parallel operation using the mapreduce library 202 c ( 464 ).
  • the executor estimates the size of the input data for each input channel of the MSCR, estimates the size of the intermediary data produced by each input channel, and estimates the size of the output data from each output channel. If any of these size estimates is equal to or exceeds 64 megabytes (MB), then the MSCR is executed as a remote, parallel operation using the mapreduce library 202 b ( 464 ). When executing the MSCR as a local, sequential operation, the executor 310 may perform the appropriate operations over the data is a sequential fashion. For example, the executor may implement in-memory for-loops to access the data and perform the appropriate operations on the data.
  • MB megabytes
  • the executor 310 may estimate the number of map worker processes and reduce worker processes needed to perform the associated processing based on the configuration of the input and output channels of the MSCR. For instance, the executor may estimate the number of map worker processes for each input channel based, for example, on an estimated or known size of the input data for each input channel and, similarly, may estimate the number of reduce worker processes based, for example, on an estimated or known amount of data to be processed by each output channel. The executor 310 may then add up the number of map worker processes and reduce worker processes and cause these worker processes to be invoked using the mapreduce library 202 b.
  • Each map worker and each reduce worker is given an index number. For example, if the MSCR includes two input channels, one with 4 map worker processes and the other with 5 map worker processes, then the 9 workers may be given an index number from 1 to 9. The same may occur for the reduce worker processes. These index numbers are used to associate a given map worker process or reduce worker process with a particular input or output channel, respectively. Continuing the foregoing example, index numbers 1-4 may be associated with the first input channel, while index numbers 5-9 may be associated with the second input channel.
  • the executor 310 also translates the MSCR into a single mapreduce operation by generating a single map function that implements the multiple map operations in the input channels of the MSCR and a single reduce function that implements the multiple reduce operations in the output channels of the MSCR.
  • the map function uses the index of the map worker processes as the basis for selecting which map operation is applied to the input. For example, an if-then statement may be included as part of the map function, with the index numbers of the map workers being the decision points for the if-then statement.
  • mapreduce library 202 b assigns a map task to a map worker process
  • the worker's associated index is passed into the map function, along with an identity of the file to be worked on.
  • the index number then dictates which map operation (parallelDo) the map function invokes on the elements in the file and, thereby, which input channel the worker implements.
  • the reduce function uses the index of the reduce worker processes as the basis for selecting which reduce operation is applied to the input of the reduce worker process.
  • the reduce worker function is assigned a reduce task, the worker's associated index is passed into the reduce function, along with an identity of the file to be worked on (which contains a single flattened stream of key-grouped inputs). The index number then dictates which reduce operation the reduce function invokes on the elements in the file and, thereby, which output channel the worker implements. If the reduce worker process implements a grouping output channel, the reduce worker process performs the CombineValues “combine” operation (if any), and then the ParallelDo “reduce” operation. If the reduce worker process implements a pass-through output channel, the reduce worker process performs an ungrouping operation that outputs key/value pairs, undoing the effect of the mapreduce library's implicit shuffle.
  • Each of the MSCR operation's input channels can emit key/value pairs to any of its R output channels.
  • input channel 2 sends one output to output channel 1 and another output to output channel 3 , and nothing to output channel 2 .
  • the mapreduce library 202 b handles the shuffle on the data output by the map worker processes and then routes the output to the correct reducer worker.
  • Each of the MSCR operation's input channels can emit key/value pairs to any of its R output channels.
  • input channel 2 sends one output to output channel 1 and another output to output channel 3 , and nothing to output channel 2 .
  • This is handled, for example, by the pipeline library 202 c by using an emitToShard(key, value, shardNum) primitive in the mapreduce library 202 b , which allows the pipeline library 202 c to designate which reduce worker process a given output of a map worker process is sent to.
  • the pipeline library 202 c may compute the range of reduce worker indices corresponding to that output channel, chooses one of them using a deterministic function, and uses the emitToShard function to send the output to the chosen reducer worker.
  • the deterministic function may include a hash on the key associated with the output values, with the result of the hash determining which of the reduce worker processes within the range of indices for the output is chosen. This may ensure that all of the data associated with a particular key is sent to the same reduce worker process.
  • the mapreduce library 202 b only directly supports writing to a single output. Moreover, in one implementation of the mapreduce library 202 b , if the reduce function's output expects key/value pairs, the keys written to this output must be the same as the keys passed in to the reduce function. In contrast, in an implementation, each MSCR output channel can write to zero, one, or several outputs, with no constraints on keys. To implement these more-flexible outputs, the reduce function may write directly to the outputs, bypassing the mapreduce library's normal output mechanism. If any of the MSCR's outputs satisfies the restrictions of a mapreduce library's output, then that output can instead be implemented using the mapreduce library's normal mechanisms.
  • the executor 310 populates the deferred objects with the appropriate data to materialize the objects ( 466 ) until all operations are completed, at which time the executor 310 returns control back over to the application 202 a ( 468 ).
  • FIG. 5 shows an example execution plan transformation that illustrates ParallelDo producer-consumer fusion and sibling fusion.
  • Graph 502 illustrates the original graph that includes ParallelDo operations A 504 , B 5 06 , C 508 , and D 510 .
  • ParallelDo operations A 504 , B 506 , C 508 , and D 510 are fused into a single ParallelDo A+B+C+D 512 to form graph 550 .
  • the new ParallelDo in graph 550 creates all the leaf outputs from the original graph 502 , plus output A. 1 514 , since output A. 1 514 is needed by some other operation Op 518 .
  • Intermediate output A. 0 516 is no longer needed and is fused away in graph 550 .
  • FIGS. 6A and 6B show an example execution plan transformation 600 that illustrates MSCR fusion.
  • Graph 601 illustrates the original graph that includes three GroupByKey operations, GBK 1 602 , GBK 2 604 , and GBK 3 606 .
  • all three GroupByKey operations 602 , 604 , 606 are related, and hence seed a single MSCR operation 652 as shown in revised graph 650 .
  • GBK 1 602 is related to GBK 2 604 because they both consume outputs of ParallelDo M 2 608 .
  • GBK 2 604 is related to GBK 3 606 because they both consume PCollection M 4 . 0 612 .
  • the PCollection M 2 . 0 is needed by later operations other than GBK 1 602 , as designated by the star.
  • the PCollection M 4 . 1 is needed by later operations other than those operations forming the MSCR operation.
  • the ParallelDos M 2 608 , M 3 614 , and M 4 612 are incorporated as MSCR input channels 616 .
  • Each of the GroupByKey 602 , 604 , 606 operations becomes a grouping output channel 620 .
  • GBK 2 's output channel incorporates the CV 2 CombineValues operation 622 and the R 2 ParallelDo operation 624 .
  • the R 3 ParallelDo 626 operation is also fused into an output channel.
  • An additional identity input channel is created for the input to GBK 1 from non-ParallelDo Op 1 .
  • Two additional pass-through output channels (shown as edges from mappers to outputs) are created for the M 2 . 0 and M 4 . 1 PCollections that are used after the MSCR operation.
  • the resulting MSCR operation 650 a has 4 input channels 616 and 5 output channels 620 .
  • FIGS. 7A-7E illustrate an example of a dataflow graph transformation performed, for example, by optimizer 306 .
  • FIG. 7A illustrates the initial parallel data pipeline 700 .
  • This pipeline takes four different input sources and writes two outputs.
  • Input 1 is processed by parallelDo( ) A 702 .
  • Input 2 is processed by parallelDo( ) B 704
  • Input 3 is processed by parallelDo( ) C 706 .
  • the results of these two operations are flattenOed 708 together and fed into parallelDo( ) D 710 .
  • Input 4 is counted using the count( ) derived operation 712 , and the result is further processed by parallelDo( ) E 714 .
  • the results of parallelDo( ) A, D, and E 702 , 710 , 714 are joined together using the join( ) 716 derived operation.
  • the result of the join( ) 716 is processed further by parallelDo( ) F 718 .
  • the results of parallelDo( ) A and F 702 and 718 are written out to external files.
  • FIG. 7B illustrates the initial dataflow graph 720 , which is constructed from calls to primitives like parallelDo( ) and flatten( ) and derived operations like count( ) and join( ) which are themselves implemented by calls to lower-level operations.
  • the count( ) call expands into ParallelDo C:Map 722 , GroupByKey C:GBK 724 , and CombineValues C:CV 726
  • the join( ) call expands into ParallelDo operations J:Tag 1 726 , J:Tag 2 728 , and J:Tag 3 730 to tag each of the N input collections, Flatten J:Fltn 732 , GroupByKey J:GBK 734 , and ParallelDo J:Untag 736 to process the results.
  • FIG. 7C shows a revised dataflow graph 738 that results from a sink flattens transformation being applied to graph 720 .
  • the Flatten operation Fltn 708 is pushed down through consuming ParallelDo operations D 710 and JTag: 2 728 .
  • FIG. 7D shows a revised dataflow graph 740 that results from a ParallelDo fusion transformation being applied to graph 738 . Both producer-consumer and sibling fusion are applied to adjacent ParallelDo operations to produce ParallelDo operations 760 , 762 , 764 , 766 , and 768 .
  • FIG. 7E shows the final, revised dataflow graph 748 that results from a MSCR fusion transformation being applied to graph 740 .
  • GroupByKey operation C:GBK 724 and surrounding ParallelDo operations (C:Map 722 and C:CV 726 ) are fused into a first MSCR operation 750 .
  • GroupByKey operations J:GBK 734 becomes the core operation of a second MSCR operation 752 and is included in a grouping output channel.
  • the second MSCR operation 752 also includes the remaining ParallelDo operations 770 , 762 , 764 , and 766 in a respective input channel, and a pass through output channel 744 .
  • the original execution plan had 16 data-parallel operations (ParallelDos, GroupByKeys, and CombineValues).
  • the final plan has two MSCR operations.
  • the functionality of the pipeline library 202 c may, additionally or alternatively, be implemented as a service that allows a client system to access the functionality over a network, such as the Internet.
  • the functionality of the pipeline library 202 c can be implemented on a server system as a Web Service with a corresponding set of Web Service Application Programming Interfaces (APIs).
  • the Web Service APIs may be implemented, for example, as a Representational State Transfer (REST)-based HTTP interface or a Simple Object Access Protocol (SOAP)-based interface.
  • REST Representational State Transfer
  • SOAP Simple Object Access Protocol
  • an interface such as a web page, may be provided to access the service over the network.
  • a user may send a program developed by the user to the service from a client system.
  • the program may include a data parallel pipeline implemented using the parallel collection classes 302 and parallel operations 304 .
  • the user may designate data for the pipeline and send a message to the service to execute the program, possibly with any arguments needed for the program.
  • the service executes the program and implements the functionality of the evaluator 306 , the optimizer 308 , and the executor 310 to implement that data parallel pipeline.
  • the service then may return any outputs of the program to the client system.
  • the user program may execute on the client system, with the program using the API to implement the data parallel pipeline using the functionality of the evaluator 306 , the optimizer 308 , and the executor 310 implemented by the service.
  • the techniques described above are not limited to any particular hardware or software configuration. Rather, they may be implemented using hardware, software, or a combination of both.
  • the methods and processes described may be implemented as computer programs that are executed on programmable computers comprising at least one processor and at least one data storage system.
  • the programs may be implemented in a high-level programming language and may also be implemented in assembly or other lower level languages, if desired.
  • Any such program will typically be stored on a computer-usable storage medium or device (e.g., CD-Rom, RAM, or magnetic disk).
  • a computer-usable storage medium or device e.g., CD-Rom, RAM, or magnetic disk.
  • the instructions of the program When read into the processor of the computer and executed, the instructions of the program cause the programmable computer to carry out the various operations described above.

Abstract

A data parallel pipeline may specify multiple parallel data objects that contain multiple elements and multiple parallel operations that operate on the parallel data objects. Based on the data parallel pipeline, a dataflow graph of deferred parallel data objects and deferred parallel operations corresponding to the data parallel pipeline may be generated and one or more graph transformations may be applied to the dataflow graph to generate a revised dataflow graph that includes one or more of the deferred parallel data objects and deferred, combined parallel data operations. The deferred, combined parallel operations may be executed to produce materialized parallel data objects corresponding to the deferred parallel data objects.

Description

    CLAIM OF PRIORITY
  • This application claims priority under to U.S. Patent Application Ser. No. 61/331,148, filed on May 4, 2010, the entire contents of which are hereby incorporated by reference.
  • TECHNICAL FIELD
  • This disclosure relates to parallel processing of data.
  • BACKGROUND
  • Large-scale data processing may include parallel processing, which generally involves performing some operation over each element of a large data set. The various operations may be chained together in a data-parallel pipeline to create an efficient mechanism for processing a data set.
  • SUMMARY
  • In one aspect, a system includes one or more processing devices and one or more storage devices. The storage devices store instructions that, when executed by the one or more processing devices, implement an application, an evaluator, an optimizer, and an executor. The application includes a data parallel pipeline. The data parallel pipeline specifies multiple parallel data objects that contain multiple elements and multiple parallel operations that operate on the parallel data objects. The evaluator is configured, based on the data parallel pipeline, to generate a dataflow graph of deferred parallel data objects and deferred parallel operations corresponding to the data parallel pipeline.
  • Deferred parallel data objects, for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object. Deferred parallel operations, for example, can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • The optimizer is configured to apply one or more graph transformations to the dataflow graph to generate a revised dataflow graph that includes one or more of the deferred parallel data objects and deferred, combined parallel data operations. The executor configured to execute the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects. Materialized parallel data objects, for example, can be data structures that include the data or elements of the parallel data object.
  • Implementations of this aspect may include one or more of the following features. For example, the deferred, combined parallel data operations may include at least one generalized mapreduce operation. The generalized mapreduce operation may include multiple, parallel map operations and multiple, parallel reduce operations and be translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations. To execute the generalized mapreduce operation, the executor may be configured to translate the combined mapreduce operation to the single mapreduce operation and execute the single mapreduce operation. To execute the single mapreduce operation, the executor may be configured to determine whether to execute the single mapreduce operation as a local, sequential operation or a remote, parallel operation. To translate the generalized mapreduce operation to the single mapreduce operation, the executor may be configured to generate a map function that includes the multiple map operations and a reducer function that includes the multiple reducer operations.
  • The multiple parallel data objects may be first class objects of a host programming language.
  • The pipeline further may include a single data object that contains a single element and the dataflow graph includes a corresponding deferred single data object. At least one of the multiple parallel operations in the pipeline may operate on the single data object and one of the multiple parallel data objects and the dataflow graph may include a corresponding deferred parallel operation that operates on a deferred single data object and a deferred parallel data object.
  • The executor may be configured to cache one or more results of the execution of the deferred, combined parallel operations for use in a future execution of the data parallel pipeline.
  • In another aspect, a method includes executing an application that includes a data parallel pipeline. The data parallel pipeline specifies multiple parallel data objects that contain multiple elements and multiple parallel operations that operate on the parallel data objects. The method further includes generating, based on the data parallel pipeline, a dataflow graph of deferred parallel data objects and deferred parallel operations corresponding to the data parallel pipeline. Deferred parallel data objects, for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object. Deferred parallel operations, for example, can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • The method also includes applying one or more graph transformations to the dataflow graph to generate a revised dataflow graph that includes one or more of the deferred parallel data objects and deferred, combined parallel data operations. In addition, the method includes executing the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects. Materialized parallel data objects, for example, can be data structures that include the data or elements of the parallel data object.
  • Implementations of this aspect may include one or more of the following features. For example, the deferred, combined parallel data operations may include at least one generalized mapreduce operation. The generalized mapreduce operation may include multiple, parallel map operations and multiple, parallel reduce operations and be translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations. Executing the generalized mapreduce operation may include translating the combined mapreduce operation to the single mapreduce operation and executing the single mapreduce operation. Executing the single mapreduce operation may include determining whether to execute the single mapreduce operation as a local, sequential operation or a remote, parallel operation. Translating the generalized mapreduce operation to the single mapreduce operation may include generating a map function that includes the multiple map operations and a reducer function that includes the multiple reducer operations. The multiple parallel data objects may be first class objects of a host programming language.
  • The pipeline further may include a single data object that contains a single element and the dataflow graph includes a corresponding deferred single data object. At least one of the multiple parallel operations in the pipeline may operate on the single data object and one of the multiple parallel data objects and the dataflow graph may include a corresponding deferred parallel operation that operates on a deferred single data object and a deferred parallel data object.
  • The method may include caching one or more results of the execution of the deferred, combined parallel operations for use in a future execution of the data parallel pipeline. In another aspect, a system includes one or more processing devices and one or more storage devices. The storage devices store instructions that, when executed by the one or more processing devices, implement an executor. The executor is configured to access a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations. Deferred parallel data objects, for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object. Deferred parallel operations, for example, can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • The executor is configured to execute the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects. Materialized parallel data objects, for example, can be data structures that include the data or elements of the parallel data object. For at least one of the deferred, combined parallel operations, the executor is configured to execute the at least one deferred, combined parallel operation by determining an estimated size of data associated with the at least one deferred, combined parallel operation; determining whether the estimated size exceeds a threshold size; if the estimated size is below a threshold size, execute the at least one deferred, combined parallel operation as a local, sequential operation; and if the estimated size exceeds a threshold size, execute the at least one deferred, combined parallel operation as remote, parallel operation.
  • Implementations of this aspect may include one or more of the following features. For example, the data associated with the at least one deferred, combined parallel operation may include one or more of input data for the at least one deferred, combined parallel operation, intermediary data produced by the at least one deferred, combined parallel operation, or output data produced by the at least one deferred, combined parallel operation. The at least one deferred, combined parallel data operation may be a generalized mapreduce operation. The generalized mapreduce operation may include multiple, parallel map operations and multiple, parallel reduce operations and be translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations. To execute the generalized mapreduce operation, the executor may be configured to translate the combined mapreduce operation to the single mapreduce operation and execute the single mapreduce operation. To execute the single mapreduce operation as a remote, parallel operation, the executor may be configured to cause the single mapreduce operation to be copied and executed on multiple, different processing modules in a datacenter. To translate the generalized mapreduce operation to the single mapreduce operation, the executor may be configured to generate a map function that includes the multiple map operations and a reducer function that includes the multiple reducer operations. To determine the estimated size, the executor may be configured to access annotations in the dataflow graph that reflect an estimate of the size of the data associated with the at least one deferred, combined parallel operation.
  • In another aspect, a method includes accessing a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations. Deferred parallel data objects, for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object. Deferred parallel operations, for example, can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • The method also includes executing the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects. Materialized parallel data objects, for example, can be data structures that include the data or elements of the parallel data object. For at least one of the deferred, combined parallel operations, executing the at least one deferred, combined parallel operation includes determining an estimated size of data associated with the at least one deferred, combined parallel operation; determining whether the estimated size exceeds a threshold size; if the estimated size is below a threshold size, execute the at least one deferred, combined parallel operation as a local, sequential operation; and if the estimated size exceeds a threshold size, execute the at least one deferred, combined parallel operation as remote, parallel operation.
  • Implementations of this aspect may include one or more of the following features. For example, the data associated with the at least one deferred, combined parallel operation may include one or more of input data for the at least one deferred, combined parallel operation, intermediary data produced by the at least one deferred, combined parallel operation, or output data produced by the at least one deferred, combined parallel operation. The at least one deferred, combined parallel data operation may be a generalized mapreduce operation. The generalized mapreduce operation may include multiple, parallel map operations and multiple, parallel reduce operations and be translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations. Executing the generalized mapreduce operation may include translating the combined mapreduce operation to the single mapreduce operation and executing the single mapreduce operation. Executing the single mapreduce operation as a remote, parallel operation may include causing the single mapreduce operation to be copied and executed on multiple, different processing modules in a datacenter. Translating the generalized mapreduce operation to the single mapreduce operation may include generating a map function that includes the multiple map operations and a reducer function that includes the multiple reducer operations. Determining the estimated size may include accessing annotations in the dataflow graph that reflect an estimate of the size of the data associated with the at least one deferred, combined parallel operation.
  • In one aspect, a system includes one or more processing devices and one or more storage devices. The storage devices store instructions that, when executed by the one or more processing devices, implement an executor. The executor is configured to access a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations. Deferred parallel data objects, for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object. Deferred parallel operations, for example, can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet been) performed on the input object.
  • At least one of the deferred, combined parallel data operation is a generalized mapreduce operation. The generalized mapreduce operation includes multiple, parallel map operations and multiple, parallel reduce operations and is translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations.
  • The executor is further configured to execute the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects. Materialized parallel data objects, for example, can be data structures that include the data or elements of the parallel data object. To execute the generalized mapreduce operation, the executor is configured to translate the combined mapreduce operation to the single mapreduce operation and execute the single mapreduce operation.
  • Implementations of this aspect may include one or more of the following features. For example, to translate the generalized mapreduce operation to the single mapreduce operation, the executor may be configured to generate a map function that includes the multiple map operations and a reduce function that includes the multiple reduce operations.
  • The executor may be configured to execute the single mapreduce operation as a remote, parallel operation. To execute the single mapreduce operation as a remote, parallel operation, the executor may be configured to cause the single mapreduce operation to be executed by multiple worker processes on multiple, different processing modules. To cause the single map reduce operation to be executed by multiple worker processes, the executor may be configured to cause multiple map worker processes to be invoked for each of the map operations, where each of the multiple map worker processes is assigned an index number. Each of the map worker processes may be configured to receive the map function that implements the multiple map operations, one or more inputs associated with one of the map operations, and the map worker process' associated index, select the map operation associated with the inputs based on the map worker process' associated index, and invoke the selected map operation on the one or more inputs.
  • To cause the single map reduce operation to be executed by multiple worker processes, the executor may be configured to cause multiple reduce worker processes to be invoked for each of the reduce operations, where each of the multiple reduce worker processes is assigned an index number. Each of the multiple reduce worker processes may be configured to receive the reduce function that implements the multiple reduce operations, one or more inputs associated with one of the reduce operations, and the reduce worker process' associated index, select the reduce operation associated with the inputs based on the worker process' associated index, and invoke the selected reduce operation on the one or more inputs.
  • In another aspect, a method includes accessing a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations. Deferred parallel data objects, for example, can be data structures that include a pointer to the parallel data operation that operates on the parallel data objects, rather than the elements stored in the parallel data object. Deferred parallel operations, for example, can be data structures that include a pointer to a parallel data object that is an input to the deferred parallel operation, a pointer to a deferred parallel object that is an output of the deferred parallel operation, and a function to be (but has not yet .
  • At least one of the deferred, combined parallel data operation is a generalized mapreduce operation. The generalized mapreduce operation includes multiple, parallel map operations and multiple, parallel reduce operations and being translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations; and
  • The method also includes executing the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects. Materialized parallel data objects, for example, can be data structures that include the data or elements of the parallel data object. Executing the generalized mapreduce operation includes translating the combined mapreduce operation to the single mapreduce operation and executing the single mapreduce operation.
  • Implementations of this aspect may include one or more of the following features. For example, translating the generalized mapreduce operation to the single mapreduce operationmay include generating a map function that includes the multiple map operations and a reduce function that includes the multiple reduce operations. Executing the single mapreduce operation may include executing the single mapreduce operation as a remote, parallel operation. Executing the single mapreduce operation as a remote, parallel operation may include causing the single mapreduce operation to be executed by multiple worker processes on multiple, different processing modules.
  • Causing the single map reduce operation to be executed by multiple worker processes may include causing the multiple map worker processes to be invoked for each of the map operations, where each of the multiple map worker processes is assigned an index number. Each of the map worker processes may be configured to receive the map function that implements the multiple map operations, one or more inputs associated with one of the map operations, and the map worker process' associated index, select the map operation associated with the inputs based on the map worker process' associated index, and invoke the selected map operation on the one or more inputs.
  • Causing the single map reduce operation to be executed by multiple worker processes may include causing multiple reduce worker processes to be invoked for each of the reduce operations, where each of the multiple reduce worker processes is assigned an index number. Each of the multiple reduce worker processes may be configured to receive the reduce function that implements the multiple reduce operations, one or more inputs associated with one of the reduce operations, and the reduce worker process' associated index, select the reduce operation associated with the inputs based on the worker process' associated index, and invoke the selected reduce operation on the one or more inputs.
  • DESCRIPTION OF DRAWINGS
  • FIG. 1 is a block diagram illustrating an example of a datacenter.
  • FIG. 2 is a block diagram of an example of a processing module.
  • FIG. 3 is a block diagram illustrating an example of a pipeline library.
  • FIG. 4A is a flow chart illustrating an example of a process that may be performed by an evaluator, an optimizer, and an executor of the pipeline library.
  • FIG. 4B is a flow chart illustrating an example of a process that may be performed by the executor of the pipeline library.
  • FIGS. 5A and 5B show an example dataflow graph transformation that illustrates ParallelDo producer-consumer fusion and sibling fusion.
  • FIGS. 6A and 6B show an example dataflow graph transformation that illustrates MSCR fusion.
  • FIGS. 7A-7E illustrate an example of a dataflow graph transformation performed to generate a final dataflow graph.
  • FIG. 8 illustrates an example of an MSCR operation with 3 input channels.
  • DETAILED DESCRIPTION
  • In general, the techniques described in this document can be applied to large-scale data processing and, in particular, to large scale data-parallel pipelines. Such large-scale processing may be performed in a distributed data processing system, such as a datacenter or a network of datacenters. For example, large-scale Internet services and the massively parallel computing infrastructure that support such services may employ warehouse-sized computing systems, made up of thousands or tens of thousands of computing nodes.
  • FIG. 1 is a block diagram illustrating an example of a datacenter 100. The datacenter 100 is used to store data, perform computational tasks, and transmit data to other systems outside of the datacenter using, for example, a network connected to the datacenter. In particular, the datacenter 100 may perform large-scale data processing on massive amounts of data.
  • The datacenter 100 includes multiple racks 102. While only two racks are shown, the datacenter 100 may have many more racks. Each rack 102 can include a frame or cabinet into which components, such as processing modules 104, are mounted. In general, each processing module 104 can include a circuit board, such as a motherboard, on which a variety of computer-related components are mounted to perform data processing. The processing modules 104 within each rack 102 are interconnected to one another through, for example, a rack switch, and the racks 102 within each datacenter 100 are also interconnected through, for example, a datacenter switch.
  • In some implementations, the processing modules 104 may each take on a role as a master or slave. The master modules control scheduling and data distribution tasks amongst themselves and the slaves. A rack can include storage (e.g., one or more network attached disks) that is shared by the one or more processing modules 104 and/or each processing module 104 may include its own storage. Additionally, or alternatively, there may be remote storage connected to the racks through a network.
  • The datacenter 100 may include dedicated optical links or other dedicated communication channels, as well as supporting hardware, such as modems, bridges, routers, switches, wireless antennas and towers, and the like. The datacenter 100 may include one or more wide area networks (WANs) as well as multiple local area networks (LANs).
  • FIG. 2 is a block diagram of an example of a processing module 200, which may be used for one or more of the processing modules 104. The processing module 200 includes memory 202, one or more processing units (CPUs) 204, and one or more network or other communication interfaces 206. These components are interconnected by one or more communication buses. In some implementations, the processing module 200 may include an input/output (I/O) interface connecting the processing module to input and output devices such as a display and a keyboard. Memory 202 may include high speed random access memory and may also include non-volatile memory, such as one or more magnetic disk storage devices. Memory 202 may include mass storage that is remotely located from the CPU 204.
  • The memory 202 stores application software 202 a, a mapreduce library 202 b, a pipeline library 202 c, and an operating system 202 d (e.g., Linux). The operating system 202 d generally includes procedures for handling various basic system services and for performing hardware dependent tasks. The application software 202 a performs large-scale data processing.
  • The libraries 202 b and 202 c provide functions and classes that may be employed by the application software 202 a to perform large-scale data processing and implement data-parallel pipelines in such large-scale data processing. The mapreduce library 202 b can support the MapReduce programming model for processing massive amounts of data in parallel. The MapReduce model is described in, for example, MapReduce: Simplified Data Processing on Large Clusters, OSDI'04: Sixth Symposium on Operating System Design and Implementation, San Francisco, Calif., December, 2004 and U.S. Pat. No. 7,650,331, both of which are incorporated by reference.
  • In general, the MapReduce model provides an abstraction to application developers for how to think about their computations. The application developers can formulate their computations according to the abstraction, which can simplify the building of programs to perform large-scale parallel-data processing. The application developers can employ the MapReduce model with or without using the mapreduce library 202 b. The mapreduce library 202 b, however, can manage many of the difficult low-level tasks. Such low-level tasks may include, for example, selecting appropriate parallel worker machines, distributing to them the program to run, managing the temporary storage and flow of intermediate data between the three phases, synchronizing the overall sequencing of the phases, and coping with transient failures of machines, networks, and software.
  • The MapReduce model generally involves breaking computations down into a mapreduce operation, which includes a single map operation and a single reduce operation. The map operation performs an operation on each of the logical records in the input to compute a set of intermediate key/value pairs. The reduce operation performs an operation on the values that share the same key to combine the values in some manner. Implicit in this model is a shuffle operation, which involves grouping all of the values with the same key.
  • The mapreduce library 202 b may implement a map phase, a shuffle phase, and a reduce phase to support computations formulated according to the MapReduce model. In some implementations, to use the mapreduce library 202 b, a user program (or another library, such as pipeline library 202 c) calls the mapreduce library 202 b, specifying information identifying the input file(s), information identifying or specifying the output files to receive output data, and two application-specific data processing operators, the map operator and the reduce operator. Generally, the map operator specifies a map function that processes the input data to produce intermediate data and the reduce operator specifies a reduce function that merges or otherwise combines the intermediate data values. The mapreduce library 202 b then employs this information to implement that map phase, the shuffle phase, and the reduce phase.
  • The map phase starts by reading a collection of values or key/value pairs from an input source, such as a text file, binary record-oriented file, or MySq1 database. Large data sets may be represented by multiple, even thousands, of files (which may be referred to as shards), and multiple file shards can be read as a single logical input source. The map phase then invokes the user-defined function, the map function or Mapper, on each element, independently and in parallel. For each input element, the user-defined function emits zero or more key/value pairs, which are the outputs of the map phase.
  • The shuffle phase takes the key/value pairs emitted by the Mappers and groups together all the key/value pairs with the same key. The shuffle phase then outputs each distinct key and a stream of all the values with that key to the next phase, the reduce phase.
  • The reduce phase takes the key-grouped data emitted by the shuffle phase and invokes the user-defined function, the reduce function or Reducer, on each distinct key-and-values group, independently and in parallel. Each Reducer invocation is passed a key and an iterator over all the values associated with that key, and emits zero or more replacement values to associate with the input key. The Reducer typically performs some kind of aggregation over all the values with a given key. For some operations, the Reducer is just the identity function. The key/value pairs emitted from all the Reducer calls are then written to an output sink, e.g., a sharded file or database.
  • To implement these phases, the mapreduce library 202 b may divide the input pieces into M pieces (for example, into 64 megabyte (MB) sized files) and start up multiple copies of the program that uses the library 202 b on a cluster of machines, such as multiple ones of the processing modules 104. One of the copies may be a master copy and the rest may be worker copies that are assigned work by the master. The master selects idle workers and assigns each one a map task or a reduce task. There are M map tasks (one for each input piece). The workers assigned to a map task use the Mapper to perform the mapping operation on the inputs to produce the intermediate results, which are divided, for example, into R sets. When the intermediate results are divided into R sets, there are R reduce tasks to assign. The workers assigned to a reduce task use the Reducer to perform the reduce operation on the intermediate values to produce the output. Once all map tasks and all reduce tasks are completed, the master returns to the user program or library employing the mapreduce library 202 b. As a result, the mapreduce operation is implemented as a set of parallel operations across a cluster of processing devices.
  • For Reducers that first combine all the values with a given key using an associative, commutative operation, a separate user-defined Combiner function can be specified to perform partial combining of values associated with a given key during the map phase. Each map worker can keep a cache of key/value pairs that have been emitted from the Mapper, and use the Combiner function to combine locally as much as possible before sending the combined key/value pairs on to the Shuffle phase. The Reducer may complete the combining step by combining values from different map workers.
  • By default, the Shuffle phase may send each key-and-values group to arbitrarily but deterministically chosen reduce worker machine, with this choice determining which output file shard will hold that key's results. Alternatively, a user defined Sharder function can be specified that selects which reduce worker machine should receive the group for a given key. A user-defined Sharder can be used to aid in load balancing. The user-defined Sharder can also be used to sort the output keys into reduce “buckets,” with all the keys of the ith reduce worker being ordered before all the keys of the ith+1st reduce worker. Coupled with the fact that each reduce worker processes keys in lexicographic order, this kind of Sharder can be used to produce sorted output.
  • The pipeline library 202 c provides functions and classes that support data-parallel pipelines and, in particular, pipelines that include chains or directed acyclic graphs of mapreduce operations. The pipeline library 202 c may help alleviate some of the burdens of implementing chains of mapreduce operations. In general, many real-world computations require a chain of mapreduce stages. While some logical computations can be expressed as a mapreduce operation, others require a sequence or graph of mapreduce operations. As the complexity of the logical computations grows, the challenge of mapping the computations into physical sequences of mapreduce operations increases. Higher-level concepts such as “count the number of occurrences” or “join tables by key” are generally hand-compiled into lower-level mapreduce operations. In addition, the user may take on the additional burdens of writing a driver program to invoke the mapreduce operations in the proper sequence, and managing the creation and deletion of intermediate files holding the data.
  • The pipeline library 202 c may obviate or reduce some of the difficulty in producing data-parallel pipelines that involve multiple mapreduce operations, as well as the need for the developer to produce additional coordination code to chain together the separate mapreduce stages in such data-parallel pipelines. The pipeline library 202 c also may obviate or reduce additional work to manage the creation and later deletion of the intermediate results in between pipeline stages. As a result, the pipeline library 202 c may help prevent the logical computation itself from becoming hidden among all the low-level coordination details, thereby making it easier for new developers to understand the computation. Moreover, making use of the pipeline library 202 c may help prevent the division of the pipeline into particular stages from becoming “baked in” to the code and difficult to change later if the logical computation needs to evolve.
  • In general, the application software 202 a may employ one or both of the libraries 202 b or 202 c. An application developer may develop application software that employs the mapreduce library 202 b to perform computations formulated as a mapreduce operation.
  • The application developer may alternatively, or additionally, employ the pipeline library 202 c when developing a data-parallel pipeline that includes multiple mapreduce operations. As discussed further below, the pipeline library 202 c may allow the developer to code the computations in a more natural manner, using the native programming language in which the pipeline library 202 c is implemented, without thinking about casting the logical computation in terms of mapreduce operations or building an ordered graph of operations. The pipeline library 202 c can formulate the logical computation in terms of multiple mapreduce operations prior to execution, and then execute the computation either by implementing the mapreduce operations itself, or interfacing with the mapreduce library 202 b to implement the mapreduce operations.
  • FIG. 3 is a block diagram illustrating an example of a pipeline library 300 that may be used to implement pipeline library 200 c. The pipeline library 300 includes one or more parallel data collection classes 302, one or more parallel operations 304, an evaluator 306, an optimizer 308, and an executor 310. In general, the parallel data collection classes 302 are used to instantiate parallel data objects that hold a collection of data, and the parallel operations 304 are used to perform parallel operations on the data held by the parallel data objects. The parallel operations 304 may be composed to implement data-parallel computations and an entire pipeline, or even multiple pipelines, can be implemented using the parallel collection classes 302 and parallel operations 304.
  • Parallel data collection classes 302 and operations 304 present a simple, high-level, uniform abstraction over many different data representations and over different execution strategies. The parallel data collection classes 302 abstract away the details of how data is represented, including whether the data is represented as an in-memory data structure, as one or more files, or as an external storage service. Similarly, parallel operations 304 abstract away their implementation strategy, such as whether an operation is implemented as a local, sequential loop, as a remote parallel invocation of the mapreduce library 202 b, as a query on a database, or as a streaming computation.
  • Rather than evaluate the parallel operations as they are traversed when the data parallel pipeline is executed, the evaluator 306 defers the evaluation of parallel operations. Instead, the evaluator 306 constructs an internal execution plan dataflow graph that contains the operations and their arguments. Once the execution plan dataflow graph for the whole logical computation is constructed, the optimizer 308 revises the execution plan, for example, by applying graph transformations that fuse or combine chains of parallel operations together into a smaller number of combined operations. The revised execution plan may include a generalized mapreduce operation that includes multiple, parallel map operations and multiple, parallel reduce operations (for example, the MapShuffleCombineReduce operation described further below), but which can be translated to a single mapreduce operation with a single map function to implement the multiple map operations and a single reduce function to implement the multiple reduce operations. The executor 310 executes the revised operations using underlying primitives (e.g., MapReduce operations). When running the execution plan, the executor 310 may choose which strategy to use to implement each operation (e.g., local sequential loop vs. remote parallel MapReduce) based in part on the size of the data being processed. The executor 310 also may place remote computations near the data on which they operate, and may perform independent operations in parallel. The executor 310 also may manage the creation and cleanup of any intermediate files needed within the computation.
  • The pipeline library 300 may be implemented in any of a number of programming languages. The following describes examples of aspects of an implementation in the Java programming language.
  • The pipeline library 300 provides a parallel data collection class referred to as a PCollection<T>, which is an immutable bag of elements of type T. A PCollection can either have a well-defined order (called a sequence), or the elements can be unordered (called a collection). Because they are less constrained, collections may be more efficient to generate and process than sequences. A PCollection<T> can be created by reading a file in one of several possible formats. For example, a text file can be read as a PCollection<String>, and a binary record-oriented file can be read as a PCollection<T>, given a specification of how to decode each binary record into an object of type T. When the pipeline library 300 is implemented using Java, a PCollection<T> may also be created from an in-memory Java Collection<T>.
  • Data sets represented by multiple file shards can be read in as a single logical PCollection. For example:
  •   PCollection<String> lines = readTextFileCollection(“/gfs/data/
      shakes/hamlet.txt”);
      PCollection<DocInfo> docInfos = readRecordFileCollection(“/gfs/
    webdocinfo/part-*”, recordsOf(DocInfo.class));
  • In this example, recordsOf( . . . ) specifies a particular way in which a DocInfo instance is encoded as a binary record. Other predefined encoding specifiers may include strings( ) for UTF-8-encoded text, ints( ) for a variable-length encoding of 32-bit integers, and pairsOf(e1,e2) for an encoding of pairs derived from the encodings of the components. Some implementations may allow users to specify their own custom encodings.
  • A second parallel data collection class 302 is PTable<K,V>, which represents an (immutable multi-map with keys of type K and values of type V. PTable<K,V> may be just an unordered bag of pairs. Some of the parallel operations 304 may apply only to PCollections of pairs, and in Java PTable<K,V> may be implemented as a subclass of PCollection<Pair<K,V>> to capture this abstraction. In another language, PTable<K,V> might be defined as a type synonym of PCollection<Pair<K,V>>.
  • The parallel data objects, such as PCollections, may be implemented as first class objects of the native language in which the library 300 is implemented. When this is the case, the objects may be manipulable like other objects in the native language. For example, the PCollections may be able to be passed into and returned from regular methods in the language, and may be able to be stored in other data structures of the language (although some implementations may prevent the PCollections from being stored in other PCollections). Also, regular control flow constructs of the native language may be able to be used to define computations involving objects, including functions, conditionals, and loops. For example, if Java is the native language:
  • Collection<PCollection<T2>> pcs =
       new ArrayList<...>( );
    for (Task task : tasks) {
       PCollection<T1> p1 = ...;
       PCollection<T2> p2;
       if (isFirstKind(task)) {
          p2 = doSomeWork(p1);
       } else {
          p2 = doSomeOtherWork(p1);
       }
       pcs.add(p2);
    }
  • Implementing the parallel data objects as first class objects in the native language of the library 300 may simplify the development of programs using the library, since the developer can use the parallel data objects in the same manner he or she would use other objects.
  • In addition to the parallel data collection classes, the pipeline library 300 can also include a single data collection class PObject<T> to support the ability to inspect the contents of PCollections during the execution of a pipeline. In contrast to a PCollection, which holds multiple elements, a PObject<T> is a container for a single object of type T (for example, a single native object (e.g., Java object) of type T) and any associated methods of PObjects are designed to operate on a single element. Like PCollections, PObjects can be either deferred or materialized (as described further below), allowing them to be computed as results of deferred operations in pipelines. Once a pipeline is run, the contents of a now-materialized PObject can be extracted using getValue( ).
  • For example, in an implementation using Java, an asSequentialCollection( ) operation can be applied to a PCollection<T> to yield a PObject<Collection<T>>, which can be inspected once the pipeline runs to read out all the elements of the computed PCollection as a regular Java in-memory Collection:
  • PTable<String,Integer> wordCounts = ...;
    PObject<Collection<Pair<String,Integer>>> result =
       wordCounts.asSequentialCollection( );
    ...
    FlumeJava.run( );
    for (Pair<String,Integer> count : result.getValue( )) {
       System.out.print(count. first + “: ” + count.second);
    }
  • As another example, the combine( ) operation (described below) applied to a PCollection<T> and a combining function over Ts yields a PObject<T> representing the fully combined result. Global sums and maxima can be computed this way.
  • The contents of PObjects also may be able to be examined within the execution of a pipeline, for example, using an operate( ) primitive provided by the pipeline library 300. The operate( ) primitive takes a list of PObjects and an argument OperateFn (which defines the operation to be performed on each PObject), and returns a list of PObjects. When evaluated, operate( ) extracts the contents of the now-materialized argument PObjects, and passes them into the argument OperateFn. The OperateFn returns a list of native objects, such as Java objects, and operate( ) wraps these native objects inside of PObjects, which are returned as the results. Using this primitive, arbitrary computations can be embedded within a pipeline and executed in deferred fashion. In other words, operations other than ParallelDo operations (described below), which operate on PCollections that contain multiple elements, can be included in the pipeline. For example, consider embedding a call to an external service that reads and writes files:
  • // Compute the URLs to crawl:
    PCollection<URL> urlsToCrawl = ...;
    // Crawl them, via an external service:
    PObject<String> fileOfUrlsToCrawl =
       urlsToCrawl.viewAsFile(TEXT);
    PObject<String> fileOfCrawledDocs =
       operate(fileOfUrlsToCrawl, new OperateFn( ) {
          String operate(String fileOfUrlsToCrawl) {
          return crawlUrls(fileOfUrlsToCrawl);
       }
    });
    PCollection<DocInfo> docInfos =
       readRecordFileCollection(fileOfCrawledDocs,
       recordsOf(DocInfo.class));
    // Use the crawled documents.
  • This example uses operations for converting between PCollections and PObjects containing file names. The viewAsFile( ) operation applied to a PCollection and a file format choice yields a PObject<String> containing the name of a temporary sharded file of the chosen format where the PCollection's contents may be found during execution of the pipeline. File-reading operations such as readRecordFileCollection( ) may be overloaded to allow reading files whose names are contained in PObjects.
  • In much the same way, the contents of PObjects can also be examined inside a DoFn (described below) by passing them in as side inputs to parallelDo( ). Normally, a DoFn performs an operation on each element of a PCollection, and just receives the PCollection as an input. In some cases, the operation on each PCollection may involve a value or other data stored in a PObject. In this case, the DoFn may receive the PCollection as an input, as normal, and a PObject as a side input. When the pipeline is run and the parallelDo( ) operation is eventually evaluated, the contents of any now-materialized PObject side inputs are extracted and provided to the user's DoFn, and then the DoFn is invoked on each element of the input PCollection to perform the defined operation on the element using the data from the PObject(s). For example:
  • PCollection<Integer> values = ...;
    PObject<Integer> pMaxValue = values.combine(MAX_INTS);
    PCollection<DocInfo> docInfos = ...;
    PCollection<Strings> results = docInfos.parallelDo(
       pMaxValue,
       new DoFn<DocInfo,String>( ) {
          private int maxValue;
          void setSideInputs(Integer maxValue) {
             this.maxValue = maxValue;
          }
          void process(DocInfo docInfo, EmitFn<String> emitFn) {
          ... use docInfo and maxValue ...
          }
       }, collectionOf(strings( )));
  • As described above, data-parallel operations 304 are invoked on parallel data objects, such as PCollections. The pipeline library 300 defines some primitive data-parallel operations, with other operations being implemented in terms of these primitives. One of the data-parallel primitives is parallelDo( ), which supports elementwise computation over an input PCollection<T> to produce a new output PCollection<S>. This operation takes as its main argument a DoFn<T, S>, a function-like object defining how to map each value in the input PCollection<T> into zero or more values to appear in the output PCollection<S>. This operation also takes an indication of the kind of PCollection or PTable to produce as a result. For example:
  • PCollection<String> words =
       lines.parallelDo(new DoFn<String,String>( ) {
          void process(String line, EmitFn<String> emitFn) {
             for (String word : splitIntoWords(line)) {
                emitFn.emit(word);
             }
          }
       }, collectionOf(strings( )));
  • In this code, collectionOf(strings( )) specifies that the parallelDo( ) operation should produce an unordered PCollection whose String elements should be encoded using UTF-8. Other options may include sequenceOf(elemEncoding) for ordered PCollections and tableOf(keyEncoding, valueEncoding) for PTables. emitFn is a call-back function passed to the user's process( . . . ) method, which should invoke emitFn.emit(outElem) for each outElem that should be added to the output PCollection. Subclasses of DoFn may be included, such as MapFn (implementing a map) and FilterFn (implementing a filter) to provide simpler interfaces in some cases.
  • The operation parallelDo( ) can be used to express both the map and reduce parts of a MapReduce operation. The library 300 also may include a version of parallelDo( ) that allows multiple output PCollections to be produced simultaneously from a single traversal of the input PCollection.
  • DoFn functions may be prevented from accessing any global mutable state of the enclosing program if DoFn functions can be distributed remotely and run in parallel. DoFn objects may be able to maintain local instance variable state, but there may be multiple DoFn replicas operating concurrently with no shared state.
  • A second primitive, groupByKey( ), converts a multimap of type PTable<K,V> (which can have many key/value pairs with the same key) into a uni-map of type PTable<K, Collection<V>> where each key maps to an unordered collection of all the values with that key. For example, the following computes a table mapping URLs to the collection of documents that link to them:
  •    PTable<URL,DocInfo> backlinks =
          docInfos.parallelDo(new DoFn<DocInfo,
          Pair<URL,DocInfo>>( ) {
             void process(DocInfo docInfo,
             EmitFn<Pair<URL,DocInfo>> emitFn)
    {            for (URL targetUrl : docInfo.getLinks( )) {
                   emitFn.emit(Pair.of(targetUrl,
                   docInfo));
                }
             }
       }, tableOf(recordsOf(URL.class), recordsOf(DocInfo.class)));
       PTable<URL, Collection<DocInfo>> referringDocInfos =
       backlinks.groupByKey( );
  • The operation groupByKey( ) corresponds to the shuffle step of MapReduce. There may also be a variant that allows specifying a sorting order for the collection of values for each key.
  • A third primitive, combineValues( ), takes an input PTable<K, Collection<V>> and an associative combining function on Vs, and returns a PTable<K,V> where each input collection of values has been combined into a single output value. For example:
  •    PTable<String,Integer> wordsWithOnes =
          words.parallelDo(
             new DoFn<String, Pair<String,Integer>>( ) {
             void process(String word,
             EmitFn<Pair<String,Integer>> emitFn) {
                emitFn.emit(Pair.of(word, 1));
             }
          }, tableOf(strings( ), ints( )));
       PTable<String,Collection<Integer>> groupedWordsWithOnes =
    wordsWithOnes.groupByKey( );
       PTable<String,Integer> wordCounts =
    groupedWordsWithOnes.combineValues(SUM_INTS);
  • The operation combineValues( ) is semantically a special case of parallelDo( ), but the associativity of the combining function allows the operation to be implemented through a combination of a MapReduce Combiner (which runs as part of each mapper) and a MapReduce Reducer (to finish the combining), which may be more efficient than doing all the combining in the reducer.
  • A fourth primitive, flatten( ) takes a list of PCollection<T>s and returns a single PCollection<T> that contains all the elements of the input PCollections. The operation flatten( ) may not actually copy the inputs, but rather just view the inputs as if the inputs were one logical PCollection.
  • A pipeline typically concludes with operations that write the final resulting PCollections to external storage. For example:
      • wordCounts.writeToRecordFileTable(“/gfs/data/shakes/hamlet-counts.records”);
  • The pipeline library 300 may include a number of other operations on PCollections that are derived in terms of the above-described primitives. These derived operations may be the same as helper functions the user could write. For example, a count( ) operation takes a PCollection<T> and returns a PTable<T,Integer> mapping each distinct element of the input PCollection to the number of times the element occurs. This function may be implemented in terms of parallelDo( ), groupByKey( ), and combineValues( ), using the same pattern as was used to compute wordCounts above. The code above can be simplified to the following:
      • PTable<String,Integer>wordCounts =words. count( );
  • Another operation, join( ), implements a join over two or more PTables sharing a common key type. When applied to a multimap PTable<K,V1> and a multimap PTable<K,V2>, join( ) returns a unimap PTable<K, Pair<Collection<V1>, Collection<V2>>> that maps each key in either of the input tables to the collection of all values with that key in the first table, and the collection of all values with that key in the second table. This resulting table can be processed further to compute a traditional inner or outer-join, but it may be more efficient to be able to manipulate the value collections directly without computing their cross-product.
  • The operation join( ) may be implemented as follows:
  • 1. Apply parallelDo( ) to each input PTable<K,Vi> to convert it into a common format of type PTable<K, TaggedUnion<V1,V2>>.
  • 2. Combine the tables using flatten( ).
  • 3. Apply groupByKey( ) to the flattened table to produce a PTable<K, Collection<TaggedUnion<V1,V2>>>.
  • 4. Apply parallelDo( ) to the key-grouped table, converting each Collection<TaggedUnion<V1,V2>> into a Pair of a Collection<V1> and a Collection<V2>.
  • Another derived operation is top( ) which takes a comparison function and a count N and returns the greatest N elements of its receiver PCollection according to the comparison function. This operation may be implemented on top of parallelDo( ), groupByKey( ), and combineValues( ).
  • The operations mentioned above to read multiple file shards as a single PCollection are derived operations too, implemented using flatten( ) and the single-file read primitives.
  • As described above, the pipeline library 300 executes parallel operations lazily, using deferred evaluation. To that end, the evaluator 306 defers the evaluation of parallel operations, and instead constructs an internal execution plan dataflow graph that contains the operations and the arguments of the operations. Each parallel data object, such as a PCollection, is represented internally either in deferred (not yet computed) or materialized (computed) state. A deferred parallel data object, for example, holds a pointer to the deferred operation that computes the parallel data object. A deferred operation, in turn, may hold references to the parallel data objects that are the arguments of the deferred operation (which may themselves be deferred or materialized) and the deferred parallel data objects that are the results of the operation. When a library operation like parallelDo( ) is called, the library 300 creates a ParallelDo deferred operation object and returns a new deferred PCollection that points to the operation. In other words, as the data parallel pipeline is executed, the evaluator 306 converts the parallel data objects and parallel operations into a directed acyclic graph of deferred (unevaluated) objects and operations. This graph may be referred to as the execution plan or execution plan dataflow graph.
  • The optimizer 308 fuses chains or subgraphs of parallel operations in the dataflow graph together into a smaller number of operations (some of which may be combined operations), which the executor 310 can then execute using an underlying primitive or other logic. The optimizer 308 may be written, for example, as a series of independent graph transformations. In one implementation, the optimizer 308 performs a series of passes over the initial execution plan that reduces the number of overall operations and groups operations, with the overall goal of producing the fewest MapShuffleCombineReduce (MSCR) operations.
  • An MSCR operation includes a combination of ParallelDo, GroupByKey, CombineValues, and Flatten operations. An MSCR operation can be mapped to and run as a single mapreduce operation. An MSCR operation has M input channels (each performing a map operation) and R output channels (each performing a shuffle, a combine, and a reduce). Each input channel m takes a PCollection<Tm> as input and performs an R-output ParallelDo “map” operation on that input to produce R outputs of type PTable<Kr,Vr>s. Each output channel R flattens its M inputs and then either (a) performs a GroupByKey “shuffle,” an optional CombineValues “combine,” and a Or-output ParallelDo “reduce” (which defaults to the identity operation), and then writes the results to Or output PCollections or (b) writes the input directly as the output. The former kind of output channel may be referred to as a “grouping” channel, while the latter kind of output channel may be referred to as a “pass-through” channel. A pass-through channel may allow the output of a mapper be a result of an MSCR operation.
  • FIG. 8 illustrates an example of an MSCR operation 800 with 3 input channels 802 a, 802 b, and 802 c. The first input channel 802 a performs a ParallelDo Ml 804 a. The second input channel 802 b performs a ParallelDo M2 804 b. The third input channel 802 c performs a ParallelDo M3 804 c. The MSCR operation includes two grouping output channels 806 a and 806 b. The first grouping output channel 806 a includes a GroupByKey GBK1 808 a, CombineValues CV1 810 a, and a reducing ParallelDo R1 812 a. Similarly, the second grouping output channel includes a GroupByKey GBK2 808 b, CombineValues CV2 810 b, and a reducing ParallelDo R2 812 b. The MSCR operation 800 also includes one pass-through output channel 814.
  • MSCR generalizes the MapReduce model by allowing multiple mappers and multiple reducers and combiners, by allowing each reducer to produce multiple outputs, by removing the requirement that the reducer must produce outputs with the same key as the reducer input, and by allowing pass-through outputs. Thus, any given MSCR may include multiple, parallel map operations that each operate on different inputs and multiple reduce operations that operate on the outputs of the map operations to produce multiple different outputs. Despite its apparent greater expressiveness, each MSCR operation can be implemented using a single mapreduce operation that includes a single map function to implement the map operations on the different inputs and a single reduce function to implement the reduce operations to produce the multiple outputs.
  • Once the execution plan is revised by the optimizer 308, the executor 310 executes the revised execution plan dataflow graph. In one implementation, the pipeline library 300 performs batch execution. In other words, the executor 310 traverses the operations in the revised execution plan in forward topological order, and executes each one in turn. Independent operations may be able to be executed simultaneously. Alternatively, incremental or continuous execution of pipelines may be implemented, where incrementally added inputs lead to quick, incremental update of outputs. Further, optimization may be performed across pipelines run by multiple users over common data sources.
  • The executor 310 executes operations other than a MSCR by performing the appropriate computations that perform the operation. MSCRs are mapped to a single mapreduce operation, which is then executed.
  • In some implementations, the executor 310 first decides whether the mapreduce operation should be run locally and sequentially, or as a remote, parallel mapreduce operation (using, for example, mapreduce library 202 b). Since there is overhead in launching a remote, parallel job, local evaluation may be used for modest-size inputs where the gain from parallel processing is outweighed by the start-up overheads. Modest-size data sets may be common during development and testing. Using local evaluation for these data sets may therefore facilitate the use of regular IDEs, debuggers, profilers, and related tools, easing the task of developing programs that include data-parallel computations.
  • If the input data set appears large (e.g., greater than or equal 64 Megabytes), the executor 310 may choose to launch a remote, parallel MapReduce operation using the mapreduce library 202 b. The executor 310 may use observations of the input data sizes and estimates of the output data sizes to automatically choose a reasonable number of parallel worker machines. Users can assist in estimating output data sizes, for example by augmenting a DoFn with a method that returns the expected ratio of output data size to input data size, based on the computation represented by that DoFn. Estimates may be refined through dynamic monitoring and feedback of observed output data sizes. Relatively more parallel workers may be allocated to jobs that have a higher ratio of CPU to I/O.
  • The executor 310 may automatically create temporary files to hold the outputs of each operation executed. Once the pipeline is completed, all of these temporary files may be automatically deleted. Alternatively, or additionally, some or all of these temporary files may be deleted as soon as they are no longer needed later in the pipeline.
  • In general, the pipeline library 300 may be designed to make building and running pipelines feel as similar as possible to running a regular program in the native language for which the pipeline library was designed. When the native language is Java, using local, sequential evaluation for modest-sized inputs is one way to do so. Another way is by automatically routing any output to System.out or System.err from within a user's DoFn, such as debugging prints, from the corresponding remote MapReduce worker to the main program's output streams. Likewise, any exceptions thrown within a DoFn running on a remote MapReduce worker are captured, sent to the main program, and rethrown.
  • The library 300 may support a cached execution mode. In this mode, rather than recompute an operation, the executor 310 first attempts to reuse the result of that operation from the previous run, if it was saved in a (internal or user-visible) file and if the executor 310 determines that the operation's result hasn't changed. An operation's result may be considered unchanged if (a) the operation's inputs haven't changed, and (b) the operation's code and captured state haven't changed. The executor 310 may perform an automatic, conservative analysis to identify when reuse of previous results is guaranteed to be safe. Caching can lead to quick edit-compile-run-debug cycles, even for pipelines that would normally take hours to run. This may reduce the amount of time required to find a bug in a late pipeline stage, fix the program, and then reexecute the revised pipeline from scratch.
  • FIG. 4A is a flow chart illustrating an example of a process 400 that may be performed by the evaluator 306, the optimizer 308, and the executor 310. Based on a data parallel pipeline that includes multiple parallel data objects and multiple parallel data operations that operate on the objects, the evaluator 306 generates a dataflow graph of deferred parallel data objects and deferred parallel operations corresponding to the data parallel pipeline (402). As described above, a deferred parallel data object is one that has not yet been computed and a deferred parallel operation is one that has not been executed. For example, as a parallel data object is encountered in the data parallel pipeline, the evaluator 306 may generate a data structure that holds a pointer to the parallel data operation that operates on the parallel data object. Similarly, as a parallel data operation is encountered, the evaluator 306 may generate a data structure that holds a pointer to a parallel data object that is an input to the deferred parallel operation and a pointer to a deferred parallel object that is an output of the deferred parallel operation.
  • Once the evaluator 306 has generated the dataflow graph, the optimizer 308 applies one or more graph transformations to the dataflow graph to generate a revised dataflow graph that includes the deferred parallel data objects (or a subset) and the deferred, combined parallel data operations (404). The deferred, combined parallel data operations may include one or more generalized mapreduce operations (for example, an MSCR), which includes multiple map operations and multiple reduce operations, but is translatable to a single mapreduce operation that includes a single map function to implement the map operations and a single reduce function to implement the reduce operations.
  • In one implementation, the optimizer 308 performs a series of passes over the dataflow graph, applying the following graph transformations or annotations in the following order: (1) sink flattens; (2) lift CombineValues operations; (3) insert fusion blocks; (4) fuse ParallelDos; and (5) fuse MSCRs.
  • The sink flattens transformation involves pushing a Flatten operation down through consuming ParallelDo operations by duplicating the ParallelDo before each input to the flatten. In other words, h(f(a)+g(b)) is equivalent to h(f(a))+h(g(b)). This transformation creates opportunities for ParallelDo fusion (described below).
  • The lift CombineValues operations annotation involves marking certain CombineValues operations for treatment as ParallelDos for ParallelDo fusion. If a CombineValues operation immediately follows a GroupByKey operation, the GroupByKey records that fact. The original CombineValues is left in place, and is henceforth treated as a normal ParallelDo operation and subject to ParallelDo fusion.
  • The insert fusion blocks annotation involves annotating the ParallelDos connecting two GroupByKey operations. If two GroupByKey operations are connected by a chain of one or more ParallelDo operations, the optimizer 308 chooses which ParallelDos should fuse up into the output channel of the earlier GroupByKey, and which should fuse down into the input channel of the later GroupByKey. The optimizer estimates the size of the intermediate PCollections along the chain of ParallelDos, identifies one with minimal expected size, and marks that intermediate PCollection as a boundary blocking ParallelDo fusion (that is, marks the ParallelDos on either side of that PCollection as not being subject to fusion into one another).
  • The fuse ParallelDos transformation involves fusing ParallelDos together. One type of ParallelDo fusion that the optimizer 306 may perform is referred to as producer-consumer fusion. If one ParallelDo operation performs function f, and the result is consumed by another ParallelDo operation that performs function g, the two ParallelDo operations may be replaced by a single ParallelDo that computes both f and g
    Figure US20140032527A1-20140130-P00001
    f If the result of the f ParallelDo is not needed by other operations in the graph, fusion has rendered it unnecessary, and the code to produce it may be removed as dead.
  • Another type of ParallelDo fusion is referred to as sibling fusion. ParallelDo sibling fusion may be applied when two or more ParallelDo operations read the same input PCollection. The ParallelDo operations can be fused into a single multi-output ParallelDo operation that computes the results of all the fused operations in a single pass over the input. Both producer-consumer and sibling fusion can apply to arbitrary trees of multi-output ParallelDo operations.
  • As mentioned earlier, CombineValues operations are special cases of ParallelDo operations that can be repeatedly applied to partially computed results. As such, ParallelDo fusion may also be applied to CombineValues operations.
  • The fuse MSCRs transformation involves creating MSCR operations. An MSCR operation starts from a set of related GroupByKey operations. GroupByKey operations may be considered related if the operations consume (possibly via Flatten operations) the same input or inputs created by the same ParallelDo operations. The MSCR's input and output channels are derived from the related GroupByKey operations and the adjacent operations in the execution plan. Each ParallelDo operation with at least one output consumed by one of the GroupByKey operations (possibly via Flatten operations) is fused into the MSCR, forming a new input channel. Any other inputs to the GroupByKeys also form new input channels with identity mappers. Each of the related GroupByKey operations starts an output channel. If a GroupByKey's result is consumed solely by a CombineValues operation, that operation is fused into the corresponding output channel. Similarly, if the GroupByKey's or fused CombineValues's result is consumed solely by a ParallelDo operation, that operation is also fused into the output channel, if it cannot be fused into a different MSCR's input channel. All the PCollections internal to the fused ParallelDo, GroupByKey, and CombineValues operations are now unnecessary and may be deleted. Finally, each output of a mapper ParallelDo that flows to an operation or output other than one of the related GroupByKeys generates its own pass-through output channel.
  • After all GroupByKey operations have been transformed into MSCR operations, any remaining ParallelDo operations are also transformed into trivial MSCR operations with a single input channel containing the ParallelDo and an identity output channel. The final optimized execution plan contains only MSCR, Flatten, and Operate operations.
  • Once the revised dataflow graph is generated, the executor 310 executes the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects (406). Executing the generalized mapreduce operation (for example, MSCR) can include translating the generalized mapreduce operation to the single mapreduce operation and executing the single mapreduce operation. Before executing the single mapreduce operation, the executor 310 may decide whether to execute the single mapreduce operation as a local, sequential operation or a remote, parallel operation and then execute the single mapreduce accordingly. For example, the executor 310 may decide based on the size of the input data set, as described above.
  • FIG. 4B is a flow chart illustrating an example of a process 450 that may be performed by the executor 310 of the pipeline library 202 c to execute the revised dataflow graph. The executor 310 accesses the revised data flow graph (452) and begins traversing the data flow graph, for example, in a forward topological manner (454). As described above, in other implementations, the executor 310 may support incremental or continuous execution of pipelines.
  • As the executor 310 encounters non-MSCR operations (456), the executor 310 executes those operations locally using logic included in the pipeline library 202 c (458). On the other hand, when the executor 310 encounters an MSCR operation (456), the executor 310 determines whether to execute the MSCR as a local, sequential operation or, instead, as a remote, parallel operation using the mapreduce library 202 b (460). For example, the executor 310 may determine an estimated size of data associated with the MSCR and determine whether the estimated size exceeds a threshold size. If the estimated size is below the threshold size, executor 310 may execute the MSCR as a local, sequential operation (462). Conversely, if the estimated size is equal to or exceeds the threshold size, the executor 310 may execute the MSCR operation as remote, parallel operation by translating the MSCR into a single mapreduce operation and executing that mapreduce operation as a remote, parallel operation using the mapreduce library 202 c (464).
  • For instance, in one implementation, the executor estimates the size of the input data for each input channel of the MSCR, estimates the size of the intermediary data produced by each input channel, and estimates the size of the output data from each output channel. If any of these size estimates is equal to or exceeds 64 megabytes (MB), then the MSCR is executed as a remote, parallel operation using the mapreduce library 202 b (464). When executing the MSCR as a local, sequential operation, the executor 310 may perform the appropriate operations over the data is a sequential fashion. For example, the executor may implement in-memory for-loops to access the data and perform the appropriate operations on the data.
  • When executing the MSCR as a remote, parallel operation using the mapreduce library 202 b, the executor 310 may estimate the number of map worker processes and reduce worker processes needed to perform the associated processing based on the configuration of the input and output channels of the MSCR. For instance, the executor may estimate the number of map worker processes for each input channel based, for example, on an estimated or known size of the input data for each input channel and, similarly, may estimate the number of reduce worker processes based, for example, on an estimated or known amount of data to be processed by each output channel. The executor 310 may then add up the number of map worker processes and reduce worker processes and cause these worker processes to be invoked using the mapreduce library 202 b.
  • Each map worker and each reduce worker is given an index number. For example, if the MSCR includes two input channels, one with 4 map worker processes and the other with 5 map worker processes, then the 9 workers may be given an index number from 1 to 9. The same may occur for the reduce worker processes. These index numbers are used to associate a given map worker process or reduce worker process with a particular input or output channel, respectively. Continuing the foregoing example, index numbers 1-4 may be associated with the first input channel, while index numbers 5-9 may be associated with the second input channel.
  • The executor 310 also translates the MSCR into a single mapreduce operation by generating a single map function that implements the multiple map operations in the input channels of the MSCR and a single reduce function that implements the multiple reduce operations in the output channels of the MSCR. The map function uses the index of the map worker processes as the basis for selecting which map operation is applied to the input. For example, an if-then statement may be included as part of the map function, with the index numbers of the map workers being the decision points for the if-then statement.
  • Thus, as the mapreduce library 202 b assigns a map task to a map worker process, the worker's associated index is passed into the map function, along with an identity of the file to be worked on. The index number then dictates which map operation (parallelDo) the map function invokes on the elements in the file and, thereby, which input channel the worker implements.
  • Similarly, the reduce function uses the index of the reduce worker processes as the basis for selecting which reduce operation is applied to the input of the reduce worker process. As a reduce worker function is assigned a reduce task, the worker's associated index is passed into the reduce function, along with an identity of the file to be worked on (which contains a single flattened stream of key-grouped inputs). The index number then dictates which reduce operation the reduce function invokes on the elements in the file and, thereby, which output channel the worker implements. If the reduce worker process implements a grouping output channel, the reduce worker process performs the CombineValues “combine” operation (if any), and then the ParallelDo “reduce” operation. If the reduce worker process implements a pass-through output channel, the reduce worker process performs an ungrouping operation that outputs key/value pairs, undoing the effect of the mapreduce library's implicit shuffle.
  • Each of the MSCR operation's input channels can emit key/value pairs to any of its R output channels. For example, input channel 2 sends one output to output channel 1 and another output to output channel 3, and nothing to output channel 2.
  • The mapreduce library 202 b handles the shuffle on the data output by the map worker processes and then routes the output to the correct reducer worker. Each of the MSCR operation's input channels can emit key/value pairs to any of its R output channels. For example, input channel 2 sends one output to output channel 1 and another output to output channel 3, and nothing to output channel 2. This is handled, for example, by the pipeline library 202 c by using an emitToShard(key, value, shardNum) primitive in the mapreduce library 202 b, which allows the pipeline library 202 c to designate which reduce worker process a given output of a map worker process is sent to. When sending an output from a given map worker process to a particular output channel, the pipeline library 202 c may compute the range of reduce worker indices corresponding to that output channel, chooses one of them using a deterministic function, and uses the emitToShard function to send the output to the chosen reducer worker. The deterministic function may include a hash on the key associated with the output values, with the result of the hash determining which of the reduce worker processes within the range of indices for the output is chosen. This may ensure that all of the data associated with a particular key is sent to the same reduce worker process.
  • In one implementation, the mapreduce library 202 b only directly supports writing to a single output. Moreover, in one implementation of the mapreduce library 202 b, if the reduce function's output expects key/value pairs, the keys written to this output must be the same as the keys passed in to the reduce function. In contrast, in an implementation, each MSCR output channel can write to zero, one, or several outputs, with no constraints on keys. To implement these more-flexible outputs, the reduce function may write directly to the outputs, bypassing the mapreduce library's normal output mechanism. If any of the MSCR's outputs satisfies the restrictions of a mapreduce library's output, then that output can instead be implemented using the mapreduce library's normal mechanisms.
  • As each of the parallel operations is evaluated, the executor 310 populates the deferred objects with the appropriate data to materialize the objects (466) until all operations are completed, at which time the executor 310 returns control back over to the application 202 a (468).
  • FIG. 5 shows an example execution plan transformation that illustrates ParallelDo producer-consumer fusion and sibling fusion. Graph 502 illustrates the original graph that includes ParallelDo operations A 504, B 5 06, C 508, and D 510. As shown, ParallelDo operations A 504, B 506, C 508, and D 510 are fused into a single ParallelDo A+B+C+D 512 to form graph 550. The new ParallelDo in graph 550 creates all the leaf outputs from the original graph 502, plus output A.1 514, since output A.1 514 is needed by some other operation Op 518. Intermediate output A.0 516 is no longer needed and is fused away in graph 550.
  • FIGS. 6A and 6B show an example execution plan transformation 600 that illustrates MSCR fusion. Graph 601 illustrates the original graph that includes three GroupByKey operations, GBK1 602, GBK2 604, and GBK3 606. In this example, all three GroupByKey operations 602, 604, 606 are related, and hence seed a single MSCR operation 652 as shown in revised graph 650. Referring to graph 601, GBK1 602 is related to GBK2 604 because they both consume outputs of ParallelDo M2 608. GBK2 604 is related to GBK3 606 because they both consume PCollection M4.0 612. The PCollection M2.0 is needed by later operations other than GBK1 602, as designated by the star. Similarly, the PCollection M4.1 is needed by later operations other than those operations forming the MSCR operation.
  • Referring to graph 650, the ParallelDos M2 608, M3 614, and M4 612 are incorporated as MSCR input channels 616. Each of the GroupByKey 602, 604, 606 operations becomes a grouping output channel 620. GBK2's output channel incorporates the CV2 CombineValues operation 622 and the R2 ParallelDo operation 624. The R3 ParallelDo 626 operation is also fused into an output channel. An additional identity input channel is created for the input to GBK1 from non-ParallelDo Op1. Two additional pass-through output channels (shown as edges from mappers to outputs) are created for the M2.0 and M4.1 PCollections that are used after the MSCR operation. The resulting MSCR operation 650 a has 4 input channels 616 and 5 output channels 620.
  • FIGS. 7A-7E illustrate an example of a dataflow graph transformation performed, for example, by optimizer 306.
  • FIG. 7A illustrates the initial parallel data pipeline 700. For simplicity, the parallel data objects are not shown. This pipeline takes four different input sources and writes two outputs. Input1 is processed by parallelDo( ) A 702. Input2 is processed by parallelDo( ) B 704, and Input3 is processed by parallelDo( ) C 706. The results of these two operations are flattenOed 708 together and fed into parallelDo( ) D 710. Input4 is counted using the count( ) derived operation 712, and the result is further processed by parallelDo( ) E 714. The results of parallelDo( ) A, D, and E 702, 710, 714 are joined together using the join( ) 716 derived operation. The result of the join( ) 716 is processed further by parallelDo( ) F 718. Finally, the results of parallelDo( ) A and F 702 and 718 are written out to external files.
  • FIG. 7B illustrates the initial dataflow graph 720, which is constructed from calls to primitives like parallelDo( ) and flatten( ) and derived operations like count( ) and join( ) which are themselves implemented by calls to lower-level operations. In this example, the count( ) call expands into ParallelDo C:Map 722, GroupByKey C:GBK 724, and CombineValues C:CV 726, and the join( ) call expands into ParallelDo operations J:Tag1 726, J:Tag2 728, and J:Tag3 730 to tag each of the N input collections, Flatten J:Fltn 732, GroupByKey J:GBK 734, and ParallelDo J:Untag 736 to process the results.
  • FIG. 7C shows a revised dataflow graph 738 that results from a sink flattens transformation being applied to graph 720. The Flatten operation Fltn 708 is pushed down through consuming ParallelDo operations D 710 and JTag:2 728.
  • FIG. 7D shows a revised dataflow graph 740 that results from a ParallelDo fusion transformation being applied to graph 738. Both producer-consumer and sibling fusion are applied to adjacent ParallelDo operations to produce ParallelDo operations 760, 762, 764, 766, and 768.
  • FIG. 7E shows the final, revised dataflow graph 748 that results from a MSCR fusion transformation being applied to graph 740. GroupByKey operation C:GBK 724 and surrounding ParallelDo operations (C:Map 722 and C:CV 726) are fused into a first MSCR operation 750. GroupByKey operations J:GBK 734 becomes the core operation of a second MSCR operation 752 and is included in a grouping output channel. The second MSCR operation 752 also includes the remaining ParallelDo operations 770, 762, 764, and 766 in a respective input channel, and a pass through output channel 744. The original execution plan had 16 data-parallel operations (ParallelDos, GroupByKeys, and CombineValues). The final plan has two MSCR operations.
  • While described as implemented as a library, the functionality of the pipeline library 202 c may, additionally or alternatively, be implemented as a service that allows a client system to access the functionality over a network, such as the Internet. For instance, the functionality of the pipeline library 202 c can be implemented on a server system as a Web Service with a corresponding set of Web Service Application Programming Interfaces (APIs). The Web Service APIs may be implemented, for example, as a Representational State Transfer (REST)-based HTTP interface or a Simple Object Access Protocol (SOAP)-based interface. Alternatively, or additionally, an interface, such as a web page, may be provided to access the service over the network.
  • Using the API or interface, a user may send a program developed by the user to the service from a client system. The program, for example, may include a data parallel pipeline implemented using the parallel collection classes 302 and parallel operations 304. Using the API or interface, the user may designate data for the pipeline and send a message to the service to execute the program, possibly with any arguments needed for the program. Once the message is received, the service executes the program and implements the functionality of the evaluator 306, the optimizer 308, and the executor 310 to implement that data parallel pipeline. The service then may return any outputs of the program to the client system. Alternatively, or additionally, the user program may execute on the client system, with the program using the API to implement the data parallel pipeline using the functionality of the evaluator 306, the optimizer 308, and the executor 310 implemented by the service.
  • The techniques described above are not limited to any particular hardware or software configuration. Rather, they may be implemented using hardware, software, or a combination of both. The methods and processes described may be implemented as computer programs that are executed on programmable computers comprising at least one processor and at least one data storage system. The programs may be implemented in a high-level programming language and may also be implemented in assembly or other lower level languages, if desired.
  • Any such program will typically be stored on a computer-usable storage medium or device (e.g., CD-Rom, RAM, or magnetic disk). When read into the processor of the computer and executed, the instructions of the program cause the programmable computer to carry out the various operations described above.
  • A number of implementations have been described. Nevertheless, it will be understood that various modifications may be. Accordingly, other implementations are within the scope of the following claims.

Claims (18)

1-12. (canceled)
13. A system comprising:
one or more processing devices;
one or more storage devices, the storage devices storing instructions that, when executed by the one or more processing devices, implement the following:
an executor configured to:
access a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations; and
execute the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects;
wherein, for at least one of the deferred, combined parallel operations, the executor is configured to execute the at least one deferred, combined parallel operation by:
determining an estimated size of data associated with the at least one deferred, combined parallel operation;
determine whether the estimated size exceeds a threshold size;
if the estimated size is below a threshold size, execute the at least one deferred, combined parallel operation as a local, sequential operation; and
if the estimated size exceeds a threshold size, execute the at least one deferred, combined parallel operation as remote, parallel operation.
14. The system of claim 13 wherein the data associated with the at least one deferred, combined parallel operation includes one or more of input data for the at least one deferred, combined parallel operation, intermediary data produced by the at least one deferred, combined parallel operation, or output data produced by the at least one deferred, combined parallel operation
15. The system of claim 13 wherein
the at least one deferred, combined parallel data operation is a generalized mapreduce operation, the generalized mapreduce operation including multiple, parallel map operations and multiple, parallel reduce operations and being translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations; and
to execute the generalized mapreduce operation, the executor is configured to translate the combined mapreduce operation to the single mapreduce operation and execute the single mapreduce operation.
16. The system of claim 14 wherein, to execute the single mapreduce operation as a remote, parallel operation, the executor is configured to cause the single mapreduce operation to be copied and executed on multiple, different processing modules in a datacenter.
17. The system of claim 14 wherein, to translate the generalized mapreduce operation to the single mapreduce operation, the executor is configured to generate a map function that includes the multiple map operations and a reducer function that includes the multiple reducer operations.
18. The system of claim 13 wherein, to determine the estimated size, the executor is configured to access annotations in the dataflow graph that reflect an estimate of the size of the data associated with the at least one deferred, combined parallel operation
19. The system of claim 13 wherein each deferred parallel data object includes a pointer to a parallel data operation that produces the parallel data object.
20. The system of claim 13 wherein each deferred parallel operation includes a pointer to a parallel data object that is an input to the deferred parallel operation and a pointer to a deferred parallel object that is an output of the deferred parallel operation.
21. The system of claim 13 wherein each materialized object includes data contained in the object.
22. A system comprising:
one or more processing devices;
one or more storage devices, the storage devices storing instructions that, when executed by the one or more processing devices, implement the following:
an executor configured to:
access a dataflow graph that includes deferred parallel data objects and deferred, combined parallel data operations, wherein at least one of the deferred, combined parallel data operation is a generalized mapreduce operation, the generalized mapreduce operation including multiple, parallel map operations and multiple, parallel reduce operations and being translatable to a single mapreduce operation that includes a single map function to implement the multiple, parallel map operations and a single reduce function to implement the multiple, parallel reduce operations; and
execute the deferred, combined parallel operations to produce materialized parallel data objects corresponding to the deferred parallel data objects;
wherein, to execute the generalized mapreduce operation, the executor is configured to translate the combined mapreduce operation to the single mapreduce operation and execute the single mapreduce operation.
23. The system of claim 22 wherein, to translate the generalized mapreduce operation to the single mapreduce operation, the executor is configured to generate a map function that includes the multiple map operations and a reduce function that includes the multiple reduce operations.
24. The system of claim 23 wherein the executor is configured to execute the single mapreduce operation as a remote, parallel operation.
25. The system of claim 24 wherein, to execute the single mapreduce operation as a remote, parallel operation, the executor is configured to cause the single mapreduce operation to be executed by multiple worker processes on multiple, different processing modules.
26. The system of claim 24 wherein, to cause the single map reduce operation to be executed by multiple worker processes, the executor is configured to cause multiple map worker processes to be invoked for each of the map operations, where each of the multiple map worker processes is assigned an index number.
27. The system of claim 26 wherein each of the map worker processes are configured to receive the map function that implements the multiple map operations, one or more inputs associated with one of the map operations, and the map worker process' associated index, select the map operation associated with the inputs based on the map worker process' associated index, and invoke the selected map operation on the one or more inputs.
28. The system of claim 24 wherein, to cause the single map reduce operation to be executed by multiple worker processes, the executor is configured to cause multiple reduce worker processes to be invoked for each of the reduce operations, where each of the multiple reduce worker processes is assigned an index number.
29. The system of claim 28 wherein each of the multiple reduce worker processes are configured to receive the reduce function that implements the multiple reduce operations, one or more inputs associated with one of the reduce operations, and the reduce worker process' associated index, select the reduce operation associated with the inputs based on the worker process' associated index, and invoke the selected reduce operation on the one or more inputs.
US14/033,145 2010-05-04 2013-09-20 Parallel processing of data Active US8959499B2 (en)

Priority Applications (8)

Application Number Priority Date Filing Date Title
US14/033,145 US8959499B2 (en) 2010-05-04 2013-09-20 Parallel processing of data
US14/622,556 US9626202B2 (en) 2010-05-04 2015-02-13 Parallel processing of data
US15/483,044 US10133592B2 (en) 2010-05-04 2017-04-10 Parallel processing of data
US16/175,925 US10338942B2 (en) 2010-05-04 2018-10-31 Parallel processing of data
US16/449,987 US10795705B2 (en) 2010-05-04 2019-06-24 Parallel processing of data
US17/009,420 US11392398B2 (en) 2010-05-04 2020-09-01 Parallel processing of data
US17/834,256 US11755351B2 (en) 2010-05-04 2022-06-07 Parallel processing of data
US18/229,450 US20230376332A1 (en) 2010-05-04 2023-08-02 Parallel Processing Of Data

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US33114810P 2010-05-04 2010-05-04
US12/794,348 US8555265B2 (en) 2010-05-04 2010-06-04 Parallel processing of data
US14/033,145 US8959499B2 (en) 2010-05-04 2013-09-20 Parallel processing of data

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US12/794,348 Division US8555265B2 (en) 2010-05-04 2010-06-04 Parallel processing of data
US12/794,348 Continuation US8555265B2 (en) 2010-05-04 2010-06-04 Parallel processing of data

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/622,556 Continuation US9626202B2 (en) 2010-05-04 2015-02-13 Parallel processing of data

Publications (2)

Publication Number Publication Date
US20140032527A1 true US20140032527A1 (en) 2014-01-30
US8959499B2 US8959499B2 (en) 2015-02-17

Family

ID=44902741

Family Applications (13)

Application Number Title Priority Date Filing Date
US12/794,348 Active 2032-02-15 US8555265B2 (en) 2010-05-04 2010-06-04 Parallel processing of data
US12/959,022 Active 2033-07-13 US8887156B2 (en) 2010-05-04 2010-12-02 Parallel processing of data
US14/033,145 Active US8959499B2 (en) 2010-05-04 2013-09-20 Parallel processing of data
US14/528,908 Active 2031-02-25 US9477502B2 (en) 2010-05-04 2014-10-30 Parallel processing of data for an untrusted application
US14/622,556 Active US9626202B2 (en) 2010-05-04 2015-02-13 Parallel processing of data
US15/278,392 Active US9678770B2 (en) 2010-05-04 2016-09-28 Parallel processing of data for an untrusted application
US15/483,044 Active US10133592B2 (en) 2010-05-04 2017-04-10 Parallel processing of data
US15/597,564 Active US9898313B2 (en) 2010-05-04 2017-05-17 Parallel processing of data for an untrusted application
US16/175,925 Active US10338942B2 (en) 2010-05-04 2018-10-31 Parallel processing of data
US16/449,987 Active US10795705B2 (en) 2010-05-04 2019-06-24 Parallel processing of data
US17/009,420 Active US11392398B2 (en) 2010-05-04 2020-09-01 Parallel processing of data
US17/834,256 Active US11755351B2 (en) 2010-05-04 2022-06-07 Parallel processing of data
US18/229,450 Pending US20230376332A1 (en) 2010-05-04 2023-08-02 Parallel Processing Of Data

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US12/794,348 Active 2032-02-15 US8555265B2 (en) 2010-05-04 2010-06-04 Parallel processing of data
US12/959,022 Active 2033-07-13 US8887156B2 (en) 2010-05-04 2010-12-02 Parallel processing of data

Family Applications After (10)

Application Number Title Priority Date Filing Date
US14/528,908 Active 2031-02-25 US9477502B2 (en) 2010-05-04 2014-10-30 Parallel processing of data for an untrusted application
US14/622,556 Active US9626202B2 (en) 2010-05-04 2015-02-13 Parallel processing of data
US15/278,392 Active US9678770B2 (en) 2010-05-04 2016-09-28 Parallel processing of data for an untrusted application
US15/483,044 Active US10133592B2 (en) 2010-05-04 2017-04-10 Parallel processing of data
US15/597,564 Active US9898313B2 (en) 2010-05-04 2017-05-17 Parallel processing of data for an untrusted application
US16/175,925 Active US10338942B2 (en) 2010-05-04 2018-10-31 Parallel processing of data
US16/449,987 Active US10795705B2 (en) 2010-05-04 2019-06-24 Parallel processing of data
US17/009,420 Active US11392398B2 (en) 2010-05-04 2020-09-01 Parallel processing of data
US17/834,256 Active US11755351B2 (en) 2010-05-04 2022-06-07 Parallel processing of data
US18/229,450 Pending US20230376332A1 (en) 2010-05-04 2023-08-02 Parallel Processing Of Data

Country Status (7)

Country Link
US (13) US8555265B2 (en)
EP (1) EP2567313B1 (en)
KR (3) KR101870320B1 (en)
CN (2) CN105279022B (en)
CA (2) CA2798266C (en)
DE (1) DE202011110864U1 (en)
WO (1) WO2011140201A1 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150149745A1 (en) * 2013-11-25 2015-05-28 Markus Eble Parallelization with controlled data sharing
US20150261824A1 (en) * 2014-03-11 2015-09-17 Wipro Limited System and method for data validation
US20150356138A1 (en) * 2014-06-06 2015-12-10 The Mathworks, Inc. Datastore mechanism for managing out-of-memory data
US20150365474A1 (en) * 2014-06-13 2015-12-17 Fujitsu Limited Computer-readable recording medium, task assignment method, and task assignment apparatus
US20160179802A1 (en) * 2014-12-18 2016-06-23 Here Global B.V. Method, apparatus and computer program product for improved storage of key-value pairs
WO2016172015A1 (en) * 2015-04-20 2016-10-27 3M Innovative Properties Company Durable low emissivity window film constructions
US20190068504A1 (en) * 2017-08-24 2019-02-28 Google Inc. Method of executing a tuple graph program across a network
US20190384577A1 (en) * 2018-06-15 2019-12-19 Sap Se Customizing operator nodes for graphical representations of data processing pipelines
US10684781B1 (en) * 2015-12-23 2020-06-16 The Mathworks, Inc. Big data read-write reduction
US10733034B2 (en) 2018-06-15 2020-08-04 Sap Se Trace messaging for distributed execution of data processing pipelines
US10866831B2 (en) 2018-06-15 2020-12-15 Sap Se Distributed execution of data processing pipelines
US10949219B2 (en) 2018-06-15 2021-03-16 Sap Se Containerized runtime environments
US10956212B1 (en) * 2019-03-08 2021-03-23 The Mathworks, Inc. Scheduler for tall-gathering algorithms that include control flow statements
US11275485B2 (en) 2018-06-15 2022-03-15 Sap Se Data processing pipeline engine
US11429355B2 (en) 2017-08-24 2022-08-30 Google Llc System of type inference for tuple graph programs

Families Citing this family (141)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9928260B2 (en) 2008-02-11 2018-03-27 Nuix Pty Ltd Systems and methods for scalable delocalized information governance
WO2009102765A2 (en) 2008-02-11 2009-08-20 Nuix North America Inc. Parallelization of electronic discovery document indexing
US9785700B2 (en) 2008-02-11 2017-10-10 Nuix Pty Ltd Systems and methods for load-balancing by secondary processors in parallelized indexing
US11048765B1 (en) 2008-06-25 2021-06-29 Richard Paiz Search engine optimizer
KR20150042866A (en) 2008-12-02 2015-04-21 아브 이니티오 테크놀로지 엘엘시 Mapping instances of a dataset within a data management system
US9665620B2 (en) 2010-01-15 2017-05-30 Ab Initio Technology Llc Managing data queries
US8918388B1 (en) * 2010-02-26 2014-12-23 Turn Inc. Custom data warehouse on top of mapreduce
US8555265B2 (en) 2010-05-04 2013-10-08 Google Inc. Parallel processing of data
US8935232B2 (en) * 2010-06-04 2015-01-13 Yale University Query execution systems and methods
US9495427B2 (en) 2010-06-04 2016-11-15 Yale University Processing of data using a database system in communication with a data processing framework
US9336263B2 (en) 2010-06-04 2016-05-10 Yale University Data loading systems and methods
US8782434B1 (en) 2010-07-15 2014-07-15 The Research Foundation For The State University Of New York System and method for validating program execution at run-time
US9489183B2 (en) * 2010-10-12 2016-11-08 Microsoft Technology Licensing, Llc Tile communication operator
AU2011323773B2 (en) * 2010-10-25 2015-07-23 Ab Initio Technology Llc Managing data set objects in a dataflow graph that represents a computer program
US9430204B2 (en) 2010-11-19 2016-08-30 Microsoft Technology Licensing, Llc Read-only communication operator
US9507568B2 (en) * 2010-12-09 2016-11-29 Microsoft Technology Licensing, Llc Nested communication operator
US9395957B2 (en) * 2010-12-22 2016-07-19 Microsoft Technology Licensing, Llc Agile communication operator
US8713039B2 (en) * 2010-12-23 2014-04-29 Microsoft Corporation Co-map communication operator
JP6138701B2 (en) * 2011-03-04 2017-05-31 富士通株式会社 Distributed calculation method and distributed calculation system
US9116955B2 (en) 2011-05-02 2015-08-25 Ab Initio Technology Llc Managing data queries
US8954967B2 (en) * 2011-05-31 2015-02-10 International Business Machines Corporation Adaptive parallel data processing
WO2012166106A1 (en) 2011-05-31 2012-12-06 Hewlett-Packard Development Company, L.P. Estimating a performance parameter of a job having map and reduce tasks after a failure
WO2013009503A2 (en) 2011-07-08 2013-01-17 Yale University Query execution systems and methods
US9747128B1 (en) * 2011-12-21 2017-08-29 EMC IP Holding Company LLC Worldwide distributed file system model
US9361079B2 (en) * 2012-01-30 2016-06-07 Nvidia Corporation Method for compiling a parallel thread execution program for general execution
US20130219394A1 (en) * 2012-02-17 2013-08-22 Kenneth Jerome GOLDMAN System and method for a map flow worker
US8949175B2 (en) * 2012-04-17 2015-02-03 Turn Inc. Meta-data driven data ingestion using MapReduce framework
US10346375B2 (en) 2012-04-26 2019-07-09 Entit Software Llc In-database parallel analytics
CN103379114B (en) 2012-04-28 2016-12-14 国际商业机器公司 For the method and apparatus protecting private data in Map Reduce system
US9165035B2 (en) * 2012-05-10 2015-10-20 Microsoft Technology Licensing, Llc Differential dataflow
US8972986B2 (en) * 2012-05-25 2015-03-03 International Business Machines Corporation Locality-aware resource allocation for cloud computing
US8984515B2 (en) 2012-05-31 2015-03-17 International Business Machines Corporation System and method for shared execution of mixed data flows
US9542462B1 (en) * 2012-06-14 2017-01-10 Google Inc. Scaling high-level statistical languages to large, distributed datasets
US9235446B2 (en) * 2012-06-22 2016-01-12 Microsoft Technology Licensing, Llc Parallel computing execution plan optimization
JP5935889B2 (en) * 2012-08-02 2016-06-15 富士通株式会社 Data processing method, information processing apparatus, and program
US20140059552A1 (en) * 2012-08-24 2014-02-27 International Business Machines Corporation Transparent efficiency for in-memory execution of map reduce job sequences
US9122873B2 (en) 2012-09-14 2015-09-01 The Research Foundation For The State University Of New York Continuous run-time validation of program execution: a practical approach
WO2014047182A2 (en) * 2012-09-21 2014-03-27 Nyse Group, Inc. High performance data streaming
US9471651B2 (en) 2012-10-08 2016-10-18 Hewlett Packard Enterprise Development Lp Adjustment of map reduce execution
US10489360B2 (en) 2012-10-17 2019-11-26 Ab Initio Technology Llc Specifying and applying rules to data
US9146830B2 (en) * 2012-10-26 2015-09-29 Jsmapreduce Corporation Hybrid local/remote infrastructure for data processing with lightweight setup, powerful debuggability, controllability, integration, and productivity features
US9832068B2 (en) 2012-12-17 2017-11-28 Microsoft Technology Licensing, Llc Reachability-based coordination for cyclic dataflow
US9536016B2 (en) * 2013-01-16 2017-01-03 Google Inc. On-disk multimap
US20140215471A1 (en) * 2013-01-28 2014-07-31 Hewlett-Packard Development Company, L.P. Creating a model relating to execution of a job on platforms
US11809506B1 (en) 2013-02-26 2023-11-07 Richard Paiz Multivariant analyzing replicating intelligent ambience evolving system
US11741090B1 (en) 2013-02-26 2023-08-29 Richard Paiz Site rank codex search patterns
US9336058B2 (en) 2013-03-14 2016-05-10 International Business Machines Corporation Automated scheduling management of MapReduce flow-graph applications
EP2979169A4 (en) * 2013-03-29 2016-11-16 Hewlett Packard Development Co Batching tuples
US9116953B2 (en) 2013-05-17 2015-08-25 Sap Se Calculation engine with dynamic partitioning of intermediate results
US9537931B2 (en) 2013-07-29 2017-01-03 Red Hat, Inc. Dynamic object oriented remote instantiation
US9852000B2 (en) * 2013-08-27 2017-12-26 Empire Technology Development Llc Consolidating operations associated with a plurality of host devices
KR20150033453A (en) * 2013-09-24 2015-04-01 주식회사 엘지씨엔에스 Method of big data processing, apparatus performing the same and storage media storing the same
US10776325B2 (en) * 2013-11-26 2020-09-15 Ab Initio Technology Llc Parallel access to data in a distributed file system
JP6454706B2 (en) 2013-12-06 2019-01-16 アビニシオ テクノロジー エルエルシー Source code conversion
CN103678235B (en) * 2013-12-09 2016-06-08 无锡市同威科技有限公司 Based on parallel pipeline network data processing device and method
US9489225B2 (en) * 2014-01-15 2016-11-08 Cisco Technology, Inc. Allocating resources for multi-phase, distributed computing jobs
US9514171B2 (en) * 2014-02-11 2016-12-06 International Business Machines Corporation Managing database clustering indices
US9563697B1 (en) 2014-02-24 2017-02-07 Amazon Technologies, Inc. Calculating differences between datasets having differing numbers of partitions
US9367366B2 (en) * 2014-03-27 2016-06-14 Nec Corporation System and methods for collaborative query processing for large scale data processing with software defined networking
US9268597B2 (en) 2014-04-01 2016-02-23 Google Inc. Incremental parallel processing of data
US9607073B2 (en) 2014-04-17 2017-03-28 Ab Initio Technology Llc Processing data from multiple sources
US9483310B2 (en) * 2014-04-29 2016-11-01 Bluedata Software, Inc. Associating cache memory with a work process
US9298590B2 (en) * 2014-06-26 2016-03-29 Google Inc. Methods and apparatuses for automated testing of streaming applications using mapreduce-like middleware
US11099841B2 (en) * 2014-06-26 2021-08-24 Sap Se Annotations for parallelization of user-defined functions with flexible partitioning
US10248688B2 (en) * 2014-06-26 2019-04-02 Sap Se Annotations for parallelization of user-defined functions with flexible partitioning
US9286044B2 (en) * 2014-06-27 2016-03-15 International Business Machines Corporation Hybrid parallelization strategies for machine learning programs on top of MapReduce
US9613127B1 (en) * 2014-06-30 2017-04-04 Quantcast Corporation Automated load-balancing of partitions in arbitrarily imbalanced distributed mapreduce computations
EP3742284A1 (en) 2014-07-18 2020-11-25 AB Initio Technology LLC Managing lineage information
US10826930B2 (en) 2014-07-22 2020-11-03 Nuix Pty Ltd Systems and methods for parallelized custom data-processing and search
CN105302536A (en) * 2014-07-31 2016-02-03 国际商业机器公司 Configuration method and apparatus for related parameters of MapReduce application
US9933918B2 (en) * 2014-09-02 2018-04-03 Ab Initio Technology Llc Specifying control and data connections in graph-based programs
JP6479966B2 (en) * 2014-09-02 2019-03-06 アビニシオ テクノロジー エルエルシー Visual definition of a subset of components in a graph-based program through user interaction
AU2015312012B2 (en) 2014-09-02 2020-02-27 Ab Initio Technology Llc Compiling graph-based program specifications
US9760406B2 (en) 2014-09-02 2017-09-12 Ab Initio Technology Llc Controlling data processing tasks
US9626393B2 (en) 2014-09-10 2017-04-18 Ab Initio Technology Llc Conditional validation rules
GB2530052A (en) 2014-09-10 2016-03-16 Ibm Outputting map-reduce jobs to an archive file
US10095654B2 (en) * 2014-09-30 2018-10-09 International Business Machines Corporation Mapping and reducing
US10353912B2 (en) 2014-10-10 2019-07-16 Salesforce.Com, Inc. Navigation of a data extraction graph of data and metadata from a data repository
US9747089B2 (en) * 2014-10-21 2017-08-29 International Business Machines Corporation Automatic conversion of sequential array-based programs to parallel map-reduce programs
US10437819B2 (en) 2014-11-14 2019-10-08 Ab Initio Technology Llc Processing queries containing a union-type operation
US9747319B2 (en) * 2014-12-31 2017-08-29 Nexenta Systems, Inc. Read-modify-write processing of chunks at the storage server level in a distributed object storage system
CN105897805B (en) * 2015-01-04 2019-12-27 伊姆西公司 Method and device for cross-layer scheduling of resources of data center with multi-layer architecture
US10417281B2 (en) 2015-02-18 2019-09-17 Ab Initio Technology Llc Querying a data source on a network
US10397313B2 (en) * 2015-03-18 2019-08-27 Nokia Of America Corporation Data stream load balancing utilizing multiple sets of servers
KR102002143B1 (en) * 2015-04-10 2019-07-19 구글 엘엘씨 Binary conversion to native client
US11200249B2 (en) 2015-04-16 2021-12-14 Nuix Limited Systems and methods for data indexing with user-side scripting
US10133827B2 (en) 2015-05-12 2018-11-20 Oracle International Corporation Automatic generation of multi-source breadth-first search from high-level graph language
US9684526B2 (en) * 2015-05-15 2017-06-20 Ab Initio Technology Llc Techniques for configuring a generic program using controls
US10614126B2 (en) 2015-05-21 2020-04-07 Oracle International Corporation Textual query editor for graph databases that performs semantic analysis using extracted information
US10127025B2 (en) 2015-07-22 2018-11-13 Oracle International Corporation Optimization techniques for high-level graph language compilers
US9575736B2 (en) * 2015-07-22 2017-02-21 Oracle International Corporation Advanced interactive command-line front-end for graph analysis systems
US10810257B2 (en) 2015-08-27 2020-10-20 Oracle International Corporation Fast processing of path-finding queries in large graph databases
US10212255B2 (en) 2015-09-28 2019-02-19 Microsoft Technology Licensing, Llc Processing requests for multi-versioned service
US9367425B1 (en) 2015-09-30 2016-06-14 Semmle Limited Disjoint-or trees for caching aggregated dependencies
US9971570B2 (en) 2015-12-15 2018-05-15 Oracle International Corporation Automated generation of memory consumption aware code
US9715373B2 (en) * 2015-12-18 2017-07-25 International Business Machines Corporation Dynamic recompilation techniques for machine learning programs
WO2017117216A1 (en) * 2015-12-29 2017-07-06 Tao Tao Systems and methods for caching task execution
US10216926B2 (en) * 2016-01-29 2019-02-26 Cisco Technology, Inc. Isolation of untrusted code in operating system without isolation capability
KR102592611B1 (en) * 2016-02-18 2023-10-23 한국전자통신연구원 Map reduce apparatus, controller for map reduce and method thereof
CN107688493B (en) * 2016-08-05 2021-06-18 阿里巴巴集团控股有限公司 Method, device and system for training deep neural network
US9977725B2 (en) 2016-08-26 2018-05-22 Cisco Technology, Inc. Automatic classification and parallel processing of untested code in a protected runtime environment
US10025566B1 (en) * 2016-10-07 2018-07-17 The Mathworks, Inc. Scheduling technique to transform dataflow graph into efficient schedule
US10613897B1 (en) * 2016-12-21 2020-04-07 Ca, Inc. Systems and methods for creating program-specific execution environments
US10769180B2 (en) 2017-02-02 2020-09-08 International Business Machines Corporation Efficient dataflow processing for objects
US10168705B2 (en) * 2017-04-06 2019-01-01 Uber Technologies, Inc. Automatic tuning of autonomous vehicle cost functions based on human driving data
US10540398B2 (en) 2017-04-24 2020-01-21 Oracle International Corporation Multi-source breadth-first search (MS-BFS) technique and graph processing system that applies it
CN107256363B (en) * 2017-06-13 2020-03-06 杭州华澜微电子股份有限公司 High-speed encryption and decryption device composed of encryption and decryption module array
US10585945B2 (en) 2017-08-01 2020-03-10 Oracle International Corporation Methods of graph-type specialization and optimization in graph algorithm DSL compilation
US10599482B2 (en) 2017-08-24 2020-03-24 Google Llc Method for intra-subgraph optimization in tuple graph programs
CN107832144B (en) * 2017-10-20 2020-07-28 南方电网科学研究院有限责任公司 Distributed parallel computing method and device
US11294943B2 (en) * 2017-12-08 2022-04-05 International Business Machines Corporation Distributed match and association of entity key-value attribute pairs
CN108197246A (en) * 2017-12-29 2018-06-22 安徽迈普德康信息科技有限公司 A kind of underground utilities electronic map management system
US11556710B2 (en) 2018-05-11 2023-01-17 International Business Machines Corporation Processing entity groups to generate analytics
US11036471B2 (en) * 2018-06-06 2021-06-15 Sap Se Data grouping for efficient parallel processing
US11277455B2 (en) 2018-06-07 2022-03-15 Mellanox Technologies, Ltd. Streaming system
CN109189477B (en) * 2018-06-27 2021-09-28 北京中科睿芯科技集团有限公司 Instruction emission control method oriented to multi-context coarse-grained data stream structure
US11003686B2 (en) 2018-07-26 2021-05-11 Roblox Corporation Addressing data skew using map-reduce
US20220043688A1 (en) * 2018-09-11 2022-02-10 Huawei Technologies Co., Ltd. Heterogeneous Scheduling for Sequential Compute Dag
US20200106828A1 (en) * 2018-10-02 2020-04-02 Mellanox Technologies, Ltd. Parallel Computation Network Device
US10795672B2 (en) 2018-10-31 2020-10-06 Oracle International Corporation Automatic generation of multi-source breadth-first search from high-level graph language for distributed graph processing systems
US10565038B1 (en) * 2018-11-27 2020-02-18 Massachusetts Institute Of Technology Method and apparatus for graph-based computing
US11379308B2 (en) 2018-12-10 2022-07-05 Zoox, Inc. Data processing pipeline failure recovery
US11625393B2 (en) 2019-02-19 2023-04-11 Mellanox Technologies, Ltd. High performance computing system
EP3699770A1 (en) 2019-02-25 2020-08-26 Mellanox Technologies TLV Ltd. Collective communication system and methods
US11093223B2 (en) 2019-07-18 2021-08-17 Ab Initio Technology Llc Automatically converting a program written in a procedural programming language into a dataflow graph and related systems and methods
US10826801B1 (en) 2019-07-31 2020-11-03 Bank Of America Corporation Multi-level data channel and inspection architectures
US11115310B2 (en) 2019-08-06 2021-09-07 Bank Of America Corporation Multi-level data channel and inspection architectures having data pipes in parallel connections
US11470046B2 (en) 2019-08-26 2022-10-11 Bank Of America Corporation Multi-level data channel and inspection architecture including security-level-based filters for diverting network traffic
US11461135B2 (en) 2019-10-25 2022-10-04 International Business Machines Corporation Dynamically modifying the parallelism of a task in a pipeline
WO2021137669A1 (en) * 2019-12-30 2021-07-08 매니코어소프트주식회사 Method for generating program for accelerator for deep learning
CN113126958B (en) * 2019-12-31 2022-07-08 思必驰科技股份有限公司 Decision scheduling customization method and system based on information flow
US11750699B2 (en) 2020-01-15 2023-09-05 Mellanox Technologies, Ltd. Small message aggregation
US11252027B2 (en) 2020-01-23 2022-02-15 Mellanox Technologies, Ltd. Network element supporting flexible data reduction operations
CN111488888B (en) * 2020-04-10 2022-12-02 腾讯科技(深圳)有限公司 Image feature extraction method and human face feature generation device
US11876885B2 (en) 2020-07-02 2024-01-16 Mellanox Technologies, Ltd. Clock queue with arming and/or self-arming features
US11848980B2 (en) * 2020-07-09 2023-12-19 Boray Data Technology Co. Ltd. Distributed pipeline configuration in a distributed computing system
US11513991B2 (en) * 2020-10-01 2022-11-29 Qualcomm Incorporated Batch operation across an interface
US11556378B2 (en) 2020-12-14 2023-01-17 Mellanox Technologies, Ltd. Offloading execution of a multi-task parameter-dependent operation to a network device
US11556452B2 (en) 2021-03-31 2023-01-17 Bank Of America Corporation System for software compiler integrity verification
CN114579190B (en) * 2022-02-17 2022-10-14 中国科学院计算机网络信息中心 Cross-center cooperative computing arrangement method and system based on pipeline mechanism
US11922237B1 (en) 2022-09-12 2024-03-05 Mellanox Technologies, Ltd. Single-step collective operations

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100205588A1 (en) * 2009-02-09 2010-08-12 Microsoft Corporation General purpose distributed data parallel computing using a high level language
US7840949B2 (en) * 2003-11-03 2010-11-23 Ramal Acquisition Corp. System and method for data transformation using dataflow graphs
US7895586B2 (en) * 2004-06-21 2011-02-22 Sanyo Electric Co., Ltd. Data flow graph processing method, reconfigurable circuit and processing apparatus
US7941794B2 (en) * 2004-08-30 2011-05-10 Sanyo Electric Co., Ltd. Data flow graph processing method and processing apparatus provided with reconfigurable circuit
US8117603B2 (en) * 2006-09-13 2012-02-14 Nec Corporation Operation synthesis system

Family Cites Families (48)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6128642A (en) * 1997-07-22 2000-10-03 At&T Corporation Load balancing based on queue length, in a network of processor stations
US7164422B1 (en) * 2000-07-28 2007-01-16 Ab Initio Software Corporation Parameterized graphs with conditional components
WO2005010632A2 (en) 2003-06-17 2005-02-03 Pact Xpp Technologies Ag Data processing device and method
US7389510B2 (en) * 2003-11-06 2008-06-17 International Business Machines Corporation Load balancing of servers in a cluster
US7590620B1 (en) * 2004-06-18 2009-09-15 Google Inc. System and method for analyzing data records
US7756919B1 (en) * 2004-06-18 2010-07-13 Google Inc. Large-scale data processing in a distributed and parallel processing enviornment
US7650331B1 (en) 2004-06-18 2010-01-19 Google Inc. System and method for efficient large-scale data processing
JP4811271B2 (en) 2004-08-25 2011-11-09 日本電気株式会社 Information communication apparatus and program execution environment control method
US20090217020A1 (en) * 2004-11-22 2009-08-27 Yourst Matt T Commit Groups for Strand-Based Computing
US8365021B2 (en) 2005-06-17 2013-01-29 Nec Corporation Information processing device comprising a plurality of domains having a plurality of processors, recovery device, program and recovery method
US7716630B2 (en) * 2005-06-27 2010-05-11 Ab Initio Technology Llc Managing parameters for graph-based computations
US7739314B2 (en) 2005-08-15 2010-06-15 Google Inc. Scalable user clustering based on set similarity
GB0517304D0 (en) * 2005-08-23 2005-10-05 Netronome Systems Inc A system and method for processing and forwarding transmitted information
US20070083735A1 (en) * 2005-08-29 2007-04-12 Glew Andrew F Hierarchical processor
US8429630B2 (en) * 2005-09-15 2013-04-23 Ca, Inc. Globally distributed utility computing cloud
US7870556B2 (en) * 2006-05-16 2011-01-11 Ab Initio Technology Llc Managing computing resources in graph-based computations
US7844959B2 (en) 2006-09-29 2010-11-30 Microsoft Corporation Runtime optimization of distributed execution graph
US8190610B2 (en) * 2006-10-05 2012-05-29 Yahoo! Inc. MapReduce for distributed database processing
US7921416B2 (en) * 2006-10-20 2011-04-05 Yahoo! Inc. Formal language and translator for parallel processing of data
CN101568900A (en) * 2006-12-22 2009-10-28 日本电气株式会社 Parallel sort device, method, and program
US20080250227A1 (en) 2007-04-04 2008-10-09 Linderman Michael D General Purpose Multiprocessor Programming Apparatus And Method
US8296743B2 (en) * 2007-12-17 2012-10-23 Intel Corporation Compiler and runtime for heterogeneous multiprocessor systems
US8537160B2 (en) 2008-03-05 2013-09-17 Microsoft Corporation Generating distributed dataflow graphs
US9058483B2 (en) 2008-05-08 2015-06-16 Google Inc. Method for validating an untrusted native code module
JP5056644B2 (en) 2008-07-18 2012-10-24 富士通セミコンダクター株式会社 Data conversion apparatus, data conversion method and program
JP4635082B2 (en) * 2008-09-30 2011-02-16 株式会社東芝 Multiprocessor system and grouping method
US7917463B2 (en) * 2008-10-10 2011-03-29 Business.Com, Inc. System and method for data warehousing and analytics on a distributed file system
US20100162230A1 (en) 2008-12-24 2010-06-24 Yahoo! Inc. Distributed computing system for large-scale data handling
US20100175049A1 (en) 2009-01-07 2010-07-08 Microsoft Corporation Scope: a structured computations optimized for parallel execution script language
US8225277B2 (en) * 2009-01-31 2012-07-17 Ted J. Biggerstaff Non-localized constraints for automated program generation
US8239847B2 (en) * 2009-03-18 2012-08-07 Microsoft Corporation General distributed reduction for data parallel computing
US8209664B2 (en) * 2009-03-18 2012-06-26 Microsoft Corporation High level programming extensions for distributed data parallel processing
US20100281078A1 (en) 2009-04-30 2010-11-04 Microsoft Corporation Distributed data reorganization for parallel execution engines
US9733914B2 (en) * 2009-06-01 2017-08-15 National Instruments Corporation Loop parallelization analyzer for data flow programs
US8478967B2 (en) * 2009-06-01 2013-07-02 National Instruments Corporation Automatically creating parallel iterative program code in a data flow program
US8365142B2 (en) * 2009-06-15 2013-01-29 Microsoft Corporation Hypergraph implementation
US9268815B2 (en) * 2009-08-20 2016-02-23 Hewlett Packard Enterprise Development Lp Map-reduce and parallel processing in databases
US8572575B2 (en) * 2009-09-14 2013-10-29 Myspace Llc Debugging a map reduce application on a cluster
US8555265B2 (en) 2010-05-04 2013-10-08 Google Inc. Parallel processing of data
WO2011151852A1 (en) * 2010-05-31 2011-12-08 Hitachi, Ltd. Data processing method and computer system
US8381015B2 (en) * 2010-06-30 2013-02-19 International Business Machines Corporation Fault tolerance for map/reduce computing
US8510284B2 (en) * 2010-12-20 2013-08-13 Microsoft Corporation Large-scale event evaluation using realtime processors
US8751639B2 (en) * 2011-04-27 2014-06-10 Rackspace Us, Inc. Event queuing and distribution system
US9450873B2 (en) * 2011-06-28 2016-09-20 Microsoft Technology Licensing, Llc Performance isolation for clouds
US9542462B1 (en) * 2012-06-14 2017-01-10 Google Inc. Scaling high-level statistical languages to large, distributed datasets
US9454571B2 (en) * 2014-06-26 2016-09-27 Sap Se Optimization of parallelization of user-defined functions with flexible partitioning
US9350384B2 (en) * 2014-09-30 2016-05-24 International Business Machines Corporation Hierarchical data compression and computation
US10417239B2 (en) * 2017-01-13 2019-09-17 International Business Machines Corporation Reducing flow delays in a data streaming application caused by lookup operations

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7840949B2 (en) * 2003-11-03 2010-11-23 Ramal Acquisition Corp. System and method for data transformation using dataflow graphs
US8429631B2 (en) * 2003-11-03 2013-04-23 Pervasive Software, Inc. System and method for data transformation using dataflow graphs
US7895586B2 (en) * 2004-06-21 2011-02-22 Sanyo Electric Co., Ltd. Data flow graph processing method, reconfigurable circuit and processing apparatus
US7941794B2 (en) * 2004-08-30 2011-05-10 Sanyo Electric Co., Ltd. Data flow graph processing method and processing apparatus provided with reconfigurable circuit
US8117603B2 (en) * 2006-09-13 2012-02-14 Nec Corporation Operation synthesis system
US20100205588A1 (en) * 2009-02-09 2010-08-12 Microsoft Corporation General purpose distributed data parallel computing using a high level language

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Al-Shabibi et al., A debugger for flow graph based parallel applications, July 2007, 7 pages. *
Liveris et al., Retiming for Synchronous Data Flow Graphs, January 2007, 6 pages. *

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150149745A1 (en) * 2013-11-25 2015-05-28 Markus Eble Parallelization with controlled data sharing
US20150261824A1 (en) * 2014-03-11 2015-09-17 Wipro Limited System and method for data validation
US9424290B2 (en) * 2014-03-11 2016-08-23 Wipro Limited System and method for data validation
US9946754B2 (en) 2014-03-11 2018-04-17 Wipro Limited System and method for data validation
US20150356138A1 (en) * 2014-06-06 2015-12-10 The Mathworks, Inc. Datastore mechanism for managing out-of-memory data
US11169993B2 (en) * 2014-06-06 2021-11-09 The Mathworks, Inc. Datastore mechanism for managing out-of-memory data
US20150365474A1 (en) * 2014-06-13 2015-12-17 Fujitsu Limited Computer-readable recording medium, task assignment method, and task assignment apparatus
US20160179802A1 (en) * 2014-12-18 2016-06-23 Here Global B.V. Method, apparatus and computer program product for improved storage of key-value pairs
US9639566B2 (en) * 2014-12-18 2017-05-02 Here Global B.V. Method, apparatus and computer program product for improved storage of key-value pairs
WO2016172015A1 (en) * 2015-04-20 2016-10-27 3M Innovative Properties Company Durable low emissivity window film constructions
US10684781B1 (en) * 2015-12-23 2020-06-16 The Mathworks, Inc. Big data read-write reduction
US10887235B2 (en) * 2017-08-24 2021-01-05 Google Llc Method of executing a tuple graph program across a network
US20190068504A1 (en) * 2017-08-24 2019-02-28 Google Inc. Method of executing a tuple graph program across a network
US11429355B2 (en) 2017-08-24 2022-08-30 Google Llc System of type inference for tuple graph programs
US10733034B2 (en) 2018-06-15 2020-08-04 Sap Se Trace messaging for distributed execution of data processing pipelines
US10747506B2 (en) * 2018-06-15 2020-08-18 Sap Se Customizing operator nodes for graphical representations of data processing pipelines
US10866831B2 (en) 2018-06-15 2020-12-15 Sap Se Distributed execution of data processing pipelines
US20190384577A1 (en) * 2018-06-15 2019-12-19 Sap Se Customizing operator nodes for graphical representations of data processing pipelines
US10949219B2 (en) 2018-06-15 2021-03-16 Sap Se Containerized runtime environments
US11275485B2 (en) 2018-06-15 2022-03-15 Sap Se Data processing pipeline engine
US10956212B1 (en) * 2019-03-08 2021-03-23 The Mathworks, Inc. Scheduler for tall-gathering algorithms that include control flow statements

Also Published As

Publication number Publication date
US11755351B2 (en) 2023-09-12
CA3014814A1 (en) 2011-11-10
US20110276789A1 (en) 2011-11-10
US20220300310A1 (en) 2022-09-22
US9626202B2 (en) 2017-04-18
US9477502B2 (en) 2016-10-25
DE202011110864U1 (en) 2017-01-16
KR101875178B1 (en) 2018-07-06
CN103109260B (en) 2015-12-02
US8555265B2 (en) 2013-10-08
CA3014814C (en) 2020-02-11
KR101870320B1 (en) 2018-06-22
US10795705B2 (en) 2020-10-06
US11392398B2 (en) 2022-07-19
EP2567313A4 (en) 2016-12-07
EP2567313B1 (en) 2019-07-17
US20190317782A1 (en) 2019-10-17
KR20180078341A (en) 2018-07-09
US9678770B2 (en) 2017-06-13
KR20170089958A (en) 2017-08-04
US20200401429A1 (en) 2020-12-24
US8887156B2 (en) 2014-11-11
US20170242715A1 (en) 2017-08-24
US20170017797A1 (en) 2017-01-19
WO2011140201A1 (en) 2011-11-10
EP2567313A1 (en) 2013-03-13
US20170249567A1 (en) 2017-08-31
US10133592B2 (en) 2018-11-20
CA2798266A1 (en) 2011-11-10
US20190065224A1 (en) 2019-02-28
CA2798266C (en) 2018-10-02
US9898313B2 (en) 2018-02-20
CN103109260A (en) 2013-05-15
US10338942B2 (en) 2019-07-02
CN105279022B (en) 2019-10-01
US20150248304A1 (en) 2015-09-03
KR101904526B1 (en) 2018-11-28
US20230376332A1 (en) 2023-11-23
US20110276962A1 (en) 2011-11-10
KR20130114577A (en) 2013-10-17
US8959499B2 (en) 2015-02-17
US20150178114A1 (en) 2015-06-25
CN105279022A (en) 2016-01-27

Similar Documents

Publication Publication Date Title
US11392398B2 (en) Parallel processing of data
Olston et al. Automatic optimization of parallel dataflow programs
WO2013153027A1 (en) Method and system for streaming processing in a map and reduce distributed computing platform
US11861331B1 (en) Scaling high-level statistical languages to large, distributed datasets
Chien Concurrent aggregates (CA): design and experience with a concurrent object-oriented language based on aggregates
Beckett et al. A strict occam design tool
Bosse Intelligent microchip networks: an agent-on-chip synthesis framework for the design of smart and robust sensor networks
Nittel et al. Conquest: CONcurrent QUEries over Space and Time
Logothetis et al. Programming bulk-incremental dataflows
Gault Improving MapReduce Performance on Clusters
Nguyen Parallelization with standard MLA, a standard ML API for hadoop and comparison with existing large-scale solutions

Legal Events

Date Code Title Description
AS Assignment

Owner name: GOOGLE INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHAMBERS, CRAIG D.;RANIWALA, ASHISH;PERRY, FRANCES J.;AND OTHERS;SIGNING DATES FROM 20100813 TO 20100823;REEL/FRAME:031720/0553

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: GOOGLE LLC, CALIFORNIA

Free format text: CHANGE OF NAME;ASSIGNOR:GOOGLE INC.;REEL/FRAME:044334/0466

Effective date: 20170929

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8