US20170019715A1 - Media production system with scheduling feature - Google Patents
Media production system with scheduling feature Download PDFInfo
- Publication number
- US20170019715A1 US20170019715A1 US15/213,126 US201615213126A US2017019715A1 US 20170019715 A1 US20170019715 A1 US 20170019715A1 US 201615213126 A US201615213126 A US 201615213126A US 2017019715 A1 US2017019715 A1 US 2017019715A1
- Authority
- US
- United States
- Prior art keywords
- event
- computing system
- scheduled
- content
- time period
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L65/00—Network arrangements, protocols or services for supporting real-time applications in data packet communication
- H04L65/60—Network streaming of media packets
- H04L65/70—Media network packetisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0486—Drag-and-drop
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
- G06F40/186—Templates
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/02—Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
- G11B27/031—Electronic editing of digitised analogue information signals, e.g. audio or video signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L51/00—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
- H04L51/07—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail characterised by the inclusion of specific contents
- H04L51/10—Multimedia information
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L51/00—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
- H04L51/52—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L65/00—Network arrangements, protocols or services for supporting real-time applications in data packet communication
- H04L65/60—Network streaming of media packets
- H04L65/61—Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio
- H04L65/611—Network streaming of media packets for supporting one-way streaming services, e.g. Internet radio for multicast or broadcast
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L65/00—Network arrangements, protocols or services for supporting real-time applications in data packet communication
- H04L65/60—Network streaming of media packets
- H04L65/75—Media network packet handling
- H04L65/762—Media network packet handling at the source
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L65/00—Network arrangements, protocols or services for supporting real-time applications in data packet communication
- H04L65/60—Network streaming of media packets
- H04L65/75—Media network packet handling
- H04L65/764—Media network packet handling at the destination
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/02—Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/10—Protocols in which an application is distributed across nodes in the network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/2866—Architectures; Arrangements
- H04L67/30—Profiles
- H04L67/306—User profiles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/52—Network services specially adapted for the location of the user terminal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
- H04N21/23418—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
- H04N21/23424—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving splicing one content stream with another content stream, e.g. for inserting or substituting an advertisement
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/235—Processing of additional data, e.g. scrambling of additional data or processing content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/258—Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
- H04N21/25808—Management of client data
- H04N21/25841—Management of client data involving the geographical location of the client
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/258—Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
- H04N21/25866—Management of end-user data
- H04N21/25875—Management of end-user data involving end-user authentication
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/262—Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/266—Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
- H04N21/2665—Gathering content from different sources, e.g. Internet and satellite
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
- H04N21/4316—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/433—Content storage operation, e.g. storage operation in response to a pause request, caching operations
- H04N21/4334—Recording operations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/435—Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/458—Scheduling content for creating a personalised stream, e.g. by combining a locally stored advertisement with an incoming stream; Updating operations, e.g. for OS modules ; time-related management operations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/47214—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for content reservation or setting reminders; for requesting event notification, e.g. of sport results or stock market
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/482—End-user interface for program selection
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/61—Network physical structure; Signal processing
- H04N21/6106—Network physical structure; Signal processing specially adapted to the downstream path of the transmission network
- H04N21/6125—Network physical structure; Signal processing specially adapted to the downstream path of the transmission network involving transmission via Internet
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/61—Network physical structure; Signal processing
- H04N21/6156—Network physical structure; Signal processing specially adapted to the upstream path of the transmission network
- H04N21/6175—Network physical structure; Signal processing specially adapted to the upstream path of the transmission network involving transmission via Internet
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8126—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8126—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
- H04N21/8133—Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts specifically related to the content, e.g. biography of the actors in a movie, detailed information about an article seen in a video program
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
Definitions
- an example method includes (i) accessing, by a first computing system, a program schedule for a media program; (ii) identifying, by the first computing system, a second computing system, wherein the identifying is based on the second computing system being associated with an event scheduled in the accessed program schedule; (iii) determining, by the first computing system, an estimated time period remaining until the event is scheduled to begin; and (iv) causing, by the first computing system, the identified second computing system to present, via a user interface of the second computing system, the determined estimated time period remaining until the event is scheduled to begin.
- an example non-transitory computer-readable medium has stored thereon program instructions that upon execution by a processor, cause performance of a first set of acts including (i) accessing, by a first computing system, a program schedule for a media program; (ii) identifying, by the first computing system, a second computing system, wherein the identifying is based on the second computing system being associated with an event scheduled in the accessed program schedule; (iii) determining, by the first computing system, an estimated time period remaining until the event is scheduled to begin; and (iv) causing, by the first computing system, the identified second computing system to present, via a user interface of the second computing system, the determined estimated time period remaining until the event is scheduled to begin.
- an example computing system configured for performing a set of acts including (i) accessing, by a first computing system, a program schedule for a media program; (ii) identifying, by the first computing system, a second computing system, wherein the identifying is based on the second computing system being associated with an event scheduled in the accessed program schedule; (iii) determining, by the first computing system, an estimated time period remaining until the event is scheduled to begin; and (iv) causing, by the first computing system, the identified second computing system to present, via a user interface of the second computing system, the determined estimated time period remaining until the event is scheduled to begin.
- FIG. 1 is a simplified block diagram of an example computing device.
- FIG. 2 is a simplified block diagram of an example video system.
- FIG. 3 is a simplified block diagram of an example video production system.
- FIG. 4A is a simplified diagram of an example frame of video content, without content overlaid thereon.
- FIG. 4B is a simplified diagram of an example frame of video content, with content overlaid thereon.
- FIG. 5 is a simplified block diagram of an example program schedule.
- FIG. 6 is a flow chart of an example method.
- a video-production system can generate video content that can serve as or be part of a video program (e.g., a news program).
- the VPS can then transmit the video content to a video-broadcast system (VBS), which in turn can transmit the video content to a first end-user device for presentation of the video content to an end-user.
- VBS video-broadcast system
- the VPS can include various components to facilitate generating video content.
- the VPS can include a video source, a DVE system, a scheduling system, and a sequencing system.
- the video source can generate video content, and can transmit the video content to the DVE system.
- the DVE system can use the video content and a DVE template to execute a DVE, which can cause the DVE system to generate new video content that is a modified version of the received video content.
- the generated video content can include the received video content with local weather content overlaid thereon.
- the scheduling system can create a program schedule, perhaps based on input received from a user (e.g., a producer or technical director) via a user interface.
- the sequencing system can process records in the program schedule, and based on the processed records, can control one or more components of the VPS, such as the video source and the DVE system, to facilitate generating video content.
- the VPS can also include a second end-user device, a content system, and a character generator.
- the second end-user device can transmit media content to the content system.
- second end-user device can capture video content (e.g., using a camera) and can transmit the captured video content to the content system.
- the content system can then forward the received media content to the character generator.
- the content system can store the media content and later can retrieve and transmit the media content to the character generator.
- the character generator can receive the media content from the content system. The character generator can then use the received media content to generate video content that includes the received media content. Further, the character generator can transmit the generated video content to the DVE system.
- the DVE system can receive the video content and can execute a DVE, which causes the DVE system to generate video content that includes the received video content.
- the generated video content can serve as or be part of a video program.
- the VPS can integrate captured video content into a video program.
- the second end-user device can transmit video content to the VPS, which the VPS can integrate into a video program.
- the second end-user device can capture video content and, at or about the same time that the video content is being captured (i.e., without significant delay), can transmit the captured video content to the content system. This is can be referred to as a live transmission. As with other events in a video program, this type of event can be scheduled in the program schedule for the video program.
- the VPS In the case where the VPS is scheduled to, at some later time, receive a live transmission of video content provided by the second end-user device and live broadcast that video content to an audience, certain issues can arise. For example, although some users (e.g., those at the site of the VPS) can have access to the relevant program schedule for this event, the user of the second end-user device may not have access to the program schedule. As a result, the end-user can become frustrated by not knowing when the user's video content will become part of a live broadcast. In addition, the user can grow tired and become unprepared when the time comes and the live broadcast is supposed to begin.
- some users e.g., those at the site of the VPS
- the user of the second end-user device may not have access to the program schedule.
- the end-user can become frustrated by not knowing when the user's video content will become part of a live broadcast.
- the user can grow tired and become unprepared when the time comes and the live broadcast is supposed to begin.
- the sequencing system can access a program schedule for a video program.
- the sequencing system can then identify the second end-user device, based on the second end-user device being associated with a later event scheduled in the accessed program schedule.
- the event can be an event where video content captured by the end-user device is live broadcast to an audience.
- the program schedule can include, as data associated with the later event, an identifier of the second end-user device and/or information such as an IP address of the second end-user device that can allow the sequencing system to communicate with the second end-user device.
- the sequencing system can then determine an estimated time period remaining until the later event is scheduled to begin.
- the sequencing system can do this in various ways. For example, for each event in a set of linear events between a current event and the later event, the sequencing system can determine a respective estimated duration of that event. Then, the sequencing system can determine that the estimated time period remaining until the later event is scheduled to begin is a sum of the determined durations.
- the sequencing system can determine the duration of each of the one or more commercial breaks and can add these one or more durations to the sum of the durations of the events as discussed above, to arrive at the estimated time period remaining until the later event is scheduled to begin.
- the sequencing system can then cause the identified second end-user device to present the determined estimated time period.
- the sequencing system 314 can cause the end-user device 304 to display “ESTIMATED TIME REMAINING UNTIL YOU ARE ON THE AIR LIVE: 2 MINUTES, 30 SECONDS.”
- the sequencing system can periodically or continuously perform one or more of the acts described in connection with this feature to allow the second end-user device to periodically or continuously display an updated estimated time period remaining until the later event is scheduled to begin.
- FIG. 1 is a simplified block diagram of an example computing device 100 .
- the computing device can be configured to perform and/or can perform one or more acts and/or functions, such as those described in this disclosure.
- the computing device 100 can include various components, such as a processor 102 , a data storage unit 104 , a communication interface 106 , and/or a user interface 108 . Each of these components can be connected to each other via a connection mechanism 110 .
- connection mechanism means a mechanism that facilitates communication between two or more components, devices, systems, or other entities.
- a connection mechanism can be a relatively simple mechanism, such as a cable or system bus, or a relatively complex mechanism, such as a packet-based communication network (e.g., the Internet).
- a connection mechanism can include a non-tangible medium (e.g., in the case where the connection is wireless).
- the processor 102 can include a general-purpose processor (e.g., a microprocessor) and/or a special-purpose processor (e.g., a digital signal processor (DSP)).
- the processor 102 can execute program instructions contained in the data storage unit 104 as discussed below.
- the data storage unit 104 can include one or more volatile, non-volatile, removable, and/or non-removable storage components, such as magnetic, optical, and/or flash storage, and/or can be integrated in whole or in part with the processor 102 . Further, the data storage unit 104 can take the form of a non-transitory computer-readable storage medium, having stored thereon program instructions (e.g., compiled or non-compiled program logic and/or machine code) that, upon execution by the processor 102 , cause the computing device 100 to perform one or more acts and/or functions, such as those described in this disclosure. These program instructions can define and/or be part of a discrete software application. In some instances, the computing device 100 can execute program instructions in response to receiving an input, such as from the communication interface 106 and/or the user interface 108 . The data storage unit 104 can also store other types of data, such as those types described in this disclosure.
- the communication interface 106 can allow the computing device 100 to connect with and/or communicate with another other entity according to one or more protocols.
- the communication interface 106 can be a wired interface, such as an Ethernet interface or a high-definition serial-digital-interface (HD-SDI).
- the communication interface 106 can be a wireless interface, such as a cellular or WI-FI interface.
- a connection can be a direct connection or an indirect connection, the latter being a connection that passes through and/or traverses one or more entities, such as a router, switcher, or other network device.
- a transmission can be a direct transmission or an indirect transmission.
- the user interface 108 can include hardware and/or software components that facilitate interaction between the computing device 100 and a user of the computing device 100 , if applicable.
- the user interface 108 can include input components such as a keyboard, a keypad, a mouse, a touch-sensitive panel, and/or a media capturing device (e.g., a microphone and/or a camera), and/or output components such as a display device (which, for example, can be combined with a touch-sensitive panel), a sound speaker, and/or a haptic feedback system.
- the computing device 100 can take various forms, such as a workstation terminal, a desktop computer, a laptop, a tablet, a mobile phone, a set-top box, and/or a television.
- FIG. 2 is a simplified block diagram of an example video system 200 .
- the video system 200 can perform various acts and/or functions related to video content, and can be implemented as a computing system.
- the term “computing system” means a system that includes at least one computing device. In some instances, a computing system can include one or more other computing systems.
- the video system 200 can include various components, such as a VPS 202 , a VBS 204 , and an end-user device 206 , each of which can be implemented as a computing system.
- the video system 200 can also include a connection mechanism 208 , which connects the VPS 202 with the VBS 204 ; and a connection mechanism 210 , which connects the VBS 204 with the end-user device 206 .
- FIG. 3 is a simplified block diagram of an example VPS 202 .
- the VPS 202 can include various components, such as a video source 302 , an end-user device 304 , a content system 306 , a character generator 308 , a digital video-effect (DVE) system 310 , a scheduling system 312 , and a sequencing system 314 , each of which can be implemented as a computing system.
- a video source 302 a video source 302
- an end-user device 304 a content system 306
- a character generator 308 e.g., a digital video-effect (DVE) system 310
- DVE digital video-effect
- the VPS 202 can also include a connection mechanism 316 , which connects the video source 302 with the sequencing system 314 ; a connection mechanism 318 , which connects the video source 302 with the DVE system 310 ; a connection mechanism 319 , which connects the end-user device 304 with the sequencing system 314 ; a connection mechanism 320 , which connects the end-user device 304 with the content system 306 ; connection mechanism 322 , which connects the content system 306 with the sequencing system 314 ; a connection mechanism 324 , which connects the content system 306 with the character generator 308 ; a connection mechanism 326 , which connects the character generator 308 with the sequencing system 314 ; a connection mechanism 328 , which connects the character generator 308 with the DVE system 310 ; a connection mechanism 330 , which connects the DVE system 310 with the sequencing system 314 ; and a connection mechanism 332 , which connects the scheduling system 312 with the sequencing system 314 .
- the video source 302 can take various forms, such as a video server, a video camera, a satellite receiver, a character generator, or a DVE system.
- a video server is the K2 server provided by Grass Valley of San Francisco, Calif.
- the character generator 308 can take various forms.
- An example character generator is the VIZ TRIO provided by Viz Rt of Bergen, Norway.
- Another example character generator is CASPAR CG developed and distributed by the Swedish Broadcasting Corporation (SVT).
- the DVE system 310 can take various forms, such as a production switcher.
- An example production switcher is the VISION OCTANE production switcher provided by Ross Video Ltd. of Iroquois, Ontario in Canada.
- the scheduling system 312 can take various forms.
- An example scheduling system is WO TRAFFIC provided by WideOrbit, Inc. of San Francisco, Calif.
- Another example scheduling system is OSI-TRAFFIC provided by Harris Corporation of Melbourne, Fla.
- the sequencing system 314 can take various forms.
- a sequencing system is sometimes referred to in the industry as a “production automation system.”
- the VBS 204 can include various components, such as a terrestrial antenna or a satellite transmitter, each of which can be implemented as a computing system.
- Each of the video-based entities described in this disclosure can include or be integrated with a corresponding audio-based entity.
- the video content described in this disclosure can include or be integrated with corresponding audio content.
- the entities can be media-based entities such as a media production system (MPS) and a media broadcast system (MBS).
- MPS media production system
- MBS media broadcast system
- the video system 200 and/or components thereof can perform various acts and/or functions. These features and related features will now be described.
- the video system 200 can perform various acts and/or functions related to video content.
- the video system 200 can receive, generate, output, and/or transmit video content that can serve as or be part of a video program (e.g., a news program).
- the act of receiving, generating, outputting, and/or transmitting video content can occur in various ways and/or according to various standards.
- the act of receiving, outputting, and/or transmitting video content can include receiving, outputting, and/or transmitting a video stream representing the video content, such as over Internet Protocol (IP) or in accordance with the high-definition serial digital interface (HD-SDI) standard.
- the act of generating content can include generating a video stream representing the video content.
- the act of receiving, generating, outputting, and/or transmitting video content can include receiving, generating, outputting, and/or transmitting an encoded or decoded version of the video content.
- the VPS 202 can perform various acts and/or functions related to video content production.
- the VPS 202 can generate and/or output video content, and can transmit the video content to another entity, such as the VBS 204 .
- the video source 302 can generate and/or output video content, and can transmit the video content to another entity, such as the DVE system 310 .
- the VPS 202 is likely to include multiple video sources and corresponding connection mechanisms, each connecting a respective one of the video sources with the DVE system 310 .
- the video source 302 can take the form of a video server.
- a video server can record and/or store video content (e.g., in the form of a file). Further, the video server can retrieve stored video content and can use the retrieved video content to generate and/or output a video stream representing the video content. This is sometimes referred to in the industry as the video server playing out the video content.
- the video server 302 can then transmit the video stream, thereby transmitting the video content, to another entity, such as the DVE system 310 .
- the end-user device 304 can perform various acts and/or functions related to media content, perhaps based on input received from a user (e.g., a field reporter) via a user interface. For example, the end-user device 304 can obtain and/or generate media content. In one example, the end-user device can to this by using a video capturing device (e.g., a camera) of the end-user device 304 to capture video content. In one use case, this can allow the end-user device 304 to capture video content of a newsworthy event, such that it can be integrated into a news program. In another example, the end-user device 304 can download media content from a media source.
- a video capturing device e.g., a camera
- the end-user device 304 can then transmit the media content to another entity, such as the content system 306 .
- the end-user device 304 can store the media content in a data storage unit (e.g., a data storage unit of the end-user device 304 ).
- the end-user device 304 can select and/or retrieve the stored media content, and can transmit it to another entity, such as the content system 306 .
- the end-user device 304 can use one software application to capture and store video content, and can then use another software application to select, retrieve, and transmit the video content.
- the end-user device 304 can capture video content and, at or about the same time that the video content is being captured (i.e., without significant delay), can transmit the captured video content to another entity, such as the content system 306 .
- This is sometimes referred to in the industry as a live video transmission or a live video stream.
- the end-user device 304 can still store, select, and/or retrieve the video content as part of this process.
- the end-user device 304 can store the video content in, and can retrieve it from, a memory buffer to facilitate the receipt and transmission of the video content.
- the end-user device 304 can use one software application to both capture and transmit video content in this manner.
- the end-user device 304 can determine data related to captured video content, which the end-user device 304 can transmit along with the video content (e.g., as metadata). For example, the end-user device 304 can determine a location of the end-user device 304 (e.g., in the form of global positioning system (GPS) coordinates) where it was capturing video content. As another example, the end-user device 304 can determine a date and/or time when the end-user device 304 was capturing video content.
- GPS global positioning system
- the content system 306 can perform various acts and/or functions related to media content, perhaps based on input received from a user (e.g., a producer or technical director) via a user interface. For example, the content system 306 can receive media content and can do so in various ways. In one example, the content system 306 can receive media content from another entity, such as the end-user device 304 .
- the content system 306 can also store, select, and/or retrieve media content. As such, the content system 306 can store received media content in a data storage unit (e.g., a data storage unit of the content system 306 ), and can then receive the media content by selecting and retrieving it from the data storage unit.
- the content system 306 can also perform similar acts in connection with data relating to media content, such as data relating to video content, as discussed above.
- the content system 306 can also modify media content and/or related data, and can do so in various ways.
- the content system 306 can modify video content using a video-editing software application. Among other things, this can allow the content system 306 to removing vulgarities, personal information, and/or extraneous information that is not suitable or desirable for integration into a video program.
- the content system can modify the related data by adding a unique identification (ID) number to the data to facilitate managing the corresponding media content.
- ID unique identification
- the content system can add a tag, keyword, or description to the corresponding media content.
- the content system 306 can also transmit media content to another entity, such as the character generator 308 .
- receiving and transmitting media content can include forwarding the media content.
- receiving and transmitting media content can include receiving the media content and transmitting a copy of the media content.
- the content system 306 can receive video content from the end-user device 304 , and can transmit a copy of the video content to the character generator 308 .
- the content system 306 can receive media content from the end-user device 304 and, at or about the same time that the media content is being received (i.e., without significant delay), can transmit the media content to another entity, such as the character generator 308 .
- the content system 306 can still store, select, and/or retrieve the media content as part of this process.
- the content system 306 can store the media content in, and can retrieve it from, a memory buffer to facilitate the receipt and transmission of the media content.
- the content system 306 can also perform various acts and/or functions related to SM content.
- SM content is content that has been published on a SM platform, which is a computer-based tool that allows users to create, share, and/or exchange content (e.g., in the form of text, images, and/or videos) in virtual communities on a computer-based network such as the Internet.
- SM platforms include TWITTER, YOUTUBE, FACEBOOK, PERISCOPE, INSTAGRAM, MEERKAT, LINKEDIN, and GOOGLE+.
- the content system 306 can receive SM content and can do so in various ways.
- the content system 306 can receive SM content by obtaining it from another entity, such as a SM platform.
- the content system 306 can obtain SM content directly from a SM platform.
- the content system 306 can obtain SM content from a SM platform via a SM dashboard application (e.g., TWEETDECK, CYFE, or HOOTSUITE).
- a SM dashboard application can provide additional searching and browsing functionalities (e.g., based on trend analysis or analytics) that may not be provided by the SM platform itself, and/or can provide access to multiple SM platforms through a single user interface.
- SM content can include various elements such as (i) data indicating the SM platform from which the SM content was received, (ii) data identifying the publisher of the SM content (e.g., an account identifier, such as a username), (iii) a profile image corresponding to the publisher of the SM content, (iv) text published by the publisher in connection with the SM content, (v) an image published by the publisher in connection with the SM content, (vi) audio content published by the publisher in connection with the SM content, (vii) video content published by the publisher in connection with the SM content (viii) a timestamp indicating a time and/or date at which the SM content was published on the SM platform, (ix) a location (e.g., represented by global positioning system (GPS) coordinates) of the publisher when the SM content was published, (x) a location at which an aspect of the SM content occurred (e.g., where video content was recorded or where a photograph was taken), (xi)
- the SM system can also store, select, and/or retrieve SM content, perhaps based on input received from a user (e.g., a producer or technical director) via a user interface.
- the content system 306 can store obtained SM content in a data storage unit (e.g., a data storage unit of the content system 306 ), and can then receive the SM content by selecting and retrieving it from the data storage unit.
- the content system 306 can select and modify SM content.
- the content system 306 can select SM content in various ways. For example, the content system 306 can select SM content responsive to the content system 306 performing an action in connection with the SM content (e.g., responsive to the content system 306 receiving or storing the SM content). In another example, the content system 306 can select SM content based on the SM content being associated with a particular characteristic (e.g., based on the SM content being scheduled to be integrated into a video program). In another example, the content system 306 can, periodically or based on a schedule, select SM content for routine processing. As yet another example, the content system 306 can select SM content based on input received from a user via a user interface.
- the content system 306 can then modify the selected SM content by identifying a first element of the selected SM content based on the first element being associated with a particular characteristic, and then modifying the selected SM content by modifying the identified first element of the selected SM content item.
- the character generator 308 can perform various acts and/or functions, perhaps based on input received via a user interface.
- the character generator 308 can receive media content and can use a character generator template and received media content to generate and/or output video content that includes the received media content.
- the content system 306 can receive media content in various ways.
- the content system 306 can receive media content from another entity.
- the content system 306 can receive video content (e.g., video content originally captured by the end-user device 304 ) from the content system 306 .
- the character generator 308 can receive SM content from the content system 306 .
- the character generator 308 can receive media content by selecting and retrieving it from a data storage unit (e.g., a data storage unit of the content system 306 ).
- the character generator template specifies the manner in which the character generator 308 uses the received media content to generate and/or output the video content.
- the character generator 308 can create and/or modify a character generator template. Further, the character generator 308 can store, select, and/or retrieve a character generator template. As such, the character generator 308 can store a character generator template in a data storage unit (e.g., a data storage unit of the character generator 308 ), and can then receive the character generator template by retrieving it from the data storage unit.
- the character generator template can specify how the character generator 308 is to receive media content. In one example, the character generator template can do so by specifying that the character generator 308 is to receive media content on a particular input of the character generator 308 (e.g., an input that maps to a particular entity, such as the content system 306 ). In another example, the character generator template can do so by specifying that the character generator 308 is to receive media content by retrieving it from a particular location of a particular data storage unit (e.g., a data storage unit of the character generator 308 ).
- a particular input of the character generator 308 e.g., an input that maps to a particular entity, such as the content system 306 .
- the character generator template can do so by specifying that the character generator 308 is to receive media content by retrieving it from a particular location of a particular data storage unit (e.g., a data storage unit of the character generator 308 ).
- the character generator 308 can use an ordered set of media content items to generate video content that includes the media content items in the specified order.
- This type of generated video content is sometimes referred to in the industry as a “ticker.”
- the media content items can include various types of content, such as text and/or images.
- the ordered set of media content items can be stored in various forms, such as in the form of an Extensible Markup Language (XML) file.
- XML Extensible Markup Language
- the character generator 308 can transmit the video content to another entity, such as the DVE system 310 , and/or can store the video content in a data storage unit (e.g., a data storage unit of the character generator 308 ).
- a data storage unit e.g., a data storage unit of the character generator 308
- the character generator 308 can receive media content, can use the media content to generate and/or output video content that includes the media content, and can transmit the video content to the DVE system 310 .
- the DVE system 310 can use a DVE template to generate and/or output video content. This is sometimes referred to in the industry as the DVE system “executing a DVE.” In some instances, the DVE system 310 can execute multiple DVEs in serial or overlapping fashion.
- the DVE template specifies the manner in which the DVE system 310 generates and/or outputs video content.
- the DVE system 310 can create and/or modify a DVE template, perhaps based on input received from a user via a user interface. Further, the DVE system 310 can store and/or retrieve a DVE template, perhaps based on input received from a user via a user interface. As such, the DVE system 310 can store a DVE system template in a data storage unit (e.g., a data storage unit of the DVE system 310 ), and can then receive the DVE template by selecting and retrieving it from the data storage unit.
- a data storage unit e.g., a data storage unit of the DVE system 310
- the DVE system 310 can use the DVE template and media content to generate and/or output video content that includes the media content.
- the DVE system 310 can receive content in various ways. For example, the DVE system 310 can do so by receiving it from another entity, such as the video source 302 and/or the character generator 308 . In another example, the DVE system 310 can do so by selecting and retrieving it from a data storage unit (e.g., a data storage unit of the DVE system 310 ).
- the DVE template can specify how the DVE system 310 is to receive media content.
- the DVE template can do so by specifying that the DVE system 310 is to receive media content on a particular input of the DVE system 310 (e.g., an input that maps to a particular entity, such as the video source 302 or the character generator 308 ).
- the DVE template can do so by specifying that the DVE system 310 is to receive content by retrieving it from a particular location of a particular data storage unit (e.g., a data storage unit of the DVE system 310 ).
- a DVE template can be configured in various ways, which can allow the DVE system 310 to execute various types of DVEs.
- a DVE template can specify that the DVE system 310 is to receive video content from the video source 302 and other media content (e.g., local weather content) from a data storage unit of the DVE system, and is to overlay the other media content on the video content, thereby generating a modified version of the video content.
- the DVE system 310 can generate video content by modifying video content.
- FIGS. 4A and 4B help illustrate this concept of overlaying other content on video content.
- FIG. 4A is a simplified depiction of an example frame 400 of video content.
- Frame 400 includes content 402 , but does not include other content overlaid on content 402 .
- FIG. 4B is a simplified depiction of another example frame 450 of video content.
- Frame 450 includes content 452 and other content 454 overlaid on content 452 .
- a DVE template can specify that the DVE system 310 is to receive first video content from the video source 302 and second video content from the character generator 308 , and is to overlay the second video content on the first video content, thereby generating a modified version of the first video content.
- a DVE template can specify that the DVE system 310 is to receive first video content from the video source 302 and second video content from the character generator 308 , and is to scale-down and re-position the first video content and the second video content, each in a respective one of two windows positioned side-by-side.
- the DVE system 310 can generate video content by scaling and/or re-positioning video content.
- the DVE system 310 can transmit the video content to another entity, such as the VBS 204 , or can store the video content in a data storage unit (e.g., a data storage unit of the DVE system 310 ).
- a data storage unit e.g., a data storage unit of the DVE system 310
- the DVE system 310 can receive first video content including video content, can use the first video content to generate and/or output second video content that includes the video content. This is an example way in which the VPS 202 can integrate video content into a video program.
- the VPS 202 can also integrate video content into a video program in other ways.
- the content system 306 can include a display device that is located within a field of the view of the video camera while the video camera records video content that serves as or is made part of the video program.
- the display device can be touch-enabled, which can allow a user (e.g., a news anchor) to interact with the video content.
- the display device and/or other components of the content system 306 can be programmed with instructions that cause particular actions in response to particular touch commands.
- the scheduling system 312 can perform various acts and/or functions related to the scheduling of video content production. For example, the scheduling system 312 can create and/or modify a program schedule of a video program, perhaps based on input received from a user via a user interface. Further, the scheduling system 312 can store and/or retrieve a program schedule, perhaps based on input received from a user via a user interface. As such, the scheduling system 312 can store a program schedule in a data storage unit (e.g., a data storage unit of the scheduling system 312 ), and can then receive the program schedule by selecting and retrieving it from the data storage unit. The scheduling system 312 can also transmit a program schedule to another entity, such as the sequencing system 314 .
- a data storage unit e.g., a data storage unit of the scheduling system 312
- the sequencing system 314 can process records in the program schedule. This can cause the sequencing system 314 to control one or more other components of the VPS 202 to facilitate the VPS 202 generating and/or outputting video content, which can serve as or be part of a video program.
- the sequencing system 314 can control the video source 302 , the content system 306 , the character generator 308 , and/or the DVE system 310 to perform the various acts and/or functions described in this disclosure.
- the sequencing system 314 can receive a program schedule in various ways. For example, the sequencing system 314 can do so by receiving it from another entity, such as the scheduling system 312 . In another example, the character generator 308 can do so by selecting and retrieving it from a data storage unit (e.g., a data storage unit of the scheduling system 312 ).
- a data storage unit e.g., a data storage unit of the scheduling system 312
- a program schedule (sometimes referred to in the industry as a “rundown”) serves as a schedule or outline of a video program.
- a program schedule can thus schedule one or more events associated with the production of a video program.
- a program schedule can include multiple records.
- a video program can be conceptually divided into multiple logically-separated portions (sometimes referred to in the industry as “stories”). As such, each portion of the video program can be represented by a separate record of the program schedule. In some cases, each record can also include one or more sub-records. Each record (including a sub-record) can include various types of data.
- FIG. 5 is a simplified diagram of an example program schedule 500 .
- the program schedule 500 includes ten records represented as ten ordered rows. Each record corresponds to a respective portion of a video program, except for one which corresponds to a commercial break. For each portion, the respective record specifies at least one data item that corresponds to that portion of the video program. In particular, each record specifies at least one of a story title, a video content item identifier, a duration, and a script.
- the duration can be an estimated duration, as discussed below.
- the script can include various instructions, such as (i) instructions for the video source 302 to playout video content, (ii) instructions for the character generator 308 to generate video content based on stored video content and/or video content received by content system 306 as a live transmission from the end-user device 304 , and forwarded to the character generator 308 , and (iii) instructions for the DVE system 310 to execute a particular DVE that overlays video content on other video content.
- the script can also include other information such as text scheduled to be displayed by a teleprompter and to be real aloud by a news anchor.
- a video content item can consist of logically-related video content.
- a video content item can be a commercial.
- a video content item can be a portion of a television program that is scheduled between two commercial breaks. This is sometimes referred to in the industry as a “program segment.”
- the first record specifies a story title of STORY A, a video content identifier of VCI ID A, a duration of 00:02:00:00 (in hours::minutes::seconds::frames format), and a script of SCRIPT A.
- SCRIPT A includes instructions for the video source 302 to playout a first video content item identified by the identifier VCI A for two minutes, instructions for the character generator 308 to use a live transmission of a second video content item provided by the end-user device 304 to generate a third video content item that includes the second video content item, and instructions for the DVE system 310 to execute a particular DVE, which causes the DVE system 310 to overlay the generated third video content item on the generated first video content item to generate a fourth video content item.
- the program schedule 500 has been greatly simplified for the purposes of illustrating certain features. In practice, a program schedule is likely to include significantly more data.
- the sequencing system 314 can process a next record (or a portion thereof) in the program schedule based on a trigger event.
- the trigger event can be the sequencing system 314 completing one or more actions related to a current record in the program schedule.
- the trigger event can be the sequencing system 314 receiving input from a user via a user interface.
- the VBS 204 can receive video content from the VPS 202 , which in turn can transmit the video content to the end-user device 206 for presentation of the video content to an end user.
- the VBS 204 can transmit video content to a large number of end-user devices for presentation of the video content to a large number of end users.
- the VBS 204 can transmit video content to the end-user device 206 in various ways.
- VBS 204 can transmit video content to the end-user device 206 over-the-air or via a packet-based network such as the Internet.
- the end-user device 206 can receive video content from the VBS 204 , and can present the video content to an end user via a user interface.
- the end-user device 304 can capture video content and, at or about the same time that the video content is being captured (i.e., without significant delay), can transmit the captured video content to another entity, such as the content system 306 . In some cases, some or all of the other operations described in this disclose can also be performed at or about that same time, which can allow the VBS 204 to transmit the captured video content to the end-user device 206 as a live transmission.
- the VPS 202 In the case where the VPS 202 is scheduled to, at some later time, receive a live transmission of video content provided by the end-user device 304 and live broadcast that video content to an audience, certain issues can arise. For example, although some users (e.g., those at the site of the VPS 202 ) can have access to the relevant program schedule for this event, the user of the end-user device 304 may not have access to the program schedule. As a result, the user can become frustrated by not knowing when the user's video content will become part of a live broadcast. In addition, the user can grow tired and become unprepared when the time comes and the live broadcast is supposed to begin.
- some users e.g., those at the site of the VPS 202
- the user of the end-user device 304 may not have access to the program schedule.
- the user can become frustrated by not knowing when the user's video content will become part of a live broadcast.
- the user can grow tired and become unprepared when the time comes and the live broadcast is supposed to begin
- VPS 202 and/or the VBS 204 can perform various acts and/or functions to help provide this feature. These acts and/or functions and related features will now be described.
- the sequencing system 314 can access a program schedule for a video program.
- the sequencing system 314 can then identify the end-user device 304 , based on the end-user device 304 being associated with a later event scheduled in the accessed program schedule.
- the event can be an event where video content captured by the end-user device 304 is live broadcast to an audience.
- the program schedule can include, as data associated with the later event, an identifier of the end-user device 304 and/or information such as an IP address of the end-user device 304 that can allow the sequencing system 314 to communicate with the end-user device 304 .
- the sequencing system 314 can include, as data associated with the later event, an identifier of a video stream (representing video content) that the end-user device 304 is already live transmitting to the VPS 202 .
- the sequencing system 314 can use mapping data (which the content system 306 can maintain for all active video streams) to map the identifier of the video stream to an identifier of the end-user device 304 .
- the sequencing system 314 can identify a video stream associated with the later event, and can use the identified video stream as a basis to identify the end-user device 304 .
- the sequencing system 314 can then determine an estimated time period remaining until the later event is scheduled to begin.
- the sequencing system 314 can do this in various ways. For example, for each event in a set of linear events between a current event and the later event, the sequencing system 314 can determine a respective estimated duration of that event. Then, the sequencing system 314 can determine that the estimated time period remaining until the later event is scheduled to begin is a sum of the determined durations. In some instances, the sequencing system 314 can also add an estimated time period remaining in the current event to this sum.
- the sequencing system 314 can estimate the duration of a given event in various ways. For example, in the case where an event has a predefined duration (e.g., where the event involves playout of a video content item having a duration of 20 seconds), the sequencing system 314 can use the predefined duration as the estimated duration. Alternatively, in the case where the duration of an event has an unknown or variable duration, the sequencing system 314 can estimate the duration of the event based on data related to the event. In some cases, the sequencing system 314 can obtain this data from the program schedule. For instance, in the case where an event involves a news anchor reading text from a teleprompter, the sequencing system 314 can use the number of words in the text to estimate how long it will take anchor to read the text.
- a predefined duration e.g., where the event involves playout of a video content item having a duration of 20 seconds
- the sequencing system 314 can estimate the duration of the event based on data related to the event. In some cases, the sequencing system 314 can obtain this data from
- the sequencing system 314 can determine the duration of each of the one or more commercial breaks and can add these one or more durations to the sum of the durations of the events as discussed above, to arrive at the estimated time period remaining until the later event is scheduled to begin. In some cases, the sequencing system 314 can determine the number and/or duration of any applicable commercial breaks based on data specified in the program schedule and/or in a corresponding broadcast automation schedule (sometimes referred to in the industry as a “traffic log”).
- the sequencing system 314 can then cause the identified end-user device 304 to present, via a user interface of the end-user device 304 , the determined estimated time period.
- the sequencing system 314 can cause the identified end-user device 304 to present the determined estimated time period in various ways. For example, the sequencing system 314 can cause this to happen by transmitting, to the end-user device 304 , an instruction that causes the end-user device 304 to present the determined estimated time period.
- the instruction can include the determined estimated time period or other data that can allow the end-user device 304 to determine the same.
- the end-user device 304 can present the determined estimated time period in various ways, such as by displaying it on a display screen of the end-user device 304 or by providing it audibly through a speaker of the end-user device 304 .
- the sequencing system 314 can cause the end-user device 304 to display “ESTIMATED TIME REMAINING UNTIL YOU ARE ON THE AIR LIVE: 2 MINUTES, 30 SECONDS.”
- the sequencing system 314 can periodically or continuously perform one or more of the acts described in connection with this feature to allow the end-user device 304 to periodically or continuously display an updated estimated time period remaining until the later event is scheduled to begin.
- the sequencing system 314 can determine an estimated time period remaining until the current event is scheduled to end.
- the sequencing system 314 can do this in various ways. For example, where the program schedule specifies a predefined or target duration, the sequencing system 314 can use this duration and a time period that has passed since the current event began to determine an estimated time period remaining until the current event is scheduled to end.
- the sequencing system 314 can then cause the identified end-user device 304 to present, via a user interface of the end-user device 304 , the estimated time period remaining until the current event is scheduled to end.
- the sequencing system 314 can cause the identified end-user device 304 to present this determined estimated time period in various ways. For example, the sequencing system 314 can cause this to happen by transmitting, to the end-user device 304 , an instruction that causes the end-user device 304 to present the determined estimated time period.
- the instruction can include the determined estimated time period or other data that can allow the end-user device 304 to determine the same.
- the end-user device 304 can present this determined estimated time period in various ways, such as by displaying it on a display screen of the end-user device 304 or by providing it audibly through a speaker of the end-user device 304 .
- the sequencing system 314 can cause the end-user device 304 to display “ESTIMATED TIME REMAINING UNTIL YOU ARE OFF THE AIR: 20 SECONDS.”
- the sequencing system 314 can periodically or continuously perform one or more of the acts described in connection with this feature to allow the end-user device 304 to periodically or continuously display an updated estimated time period remaining until the current event is scheduled to end.
- the acts and/or functions described in this disclosure have been described in the context of the video system 200 and as video-related concepts (e.g., generating a video program), the acts and/or functions can also be applied in the context of an audio system and used as audio-related concepts (e.g., generating an audio program). More generally, the acts and/or functions described in this disclosure can be applied in the context of a media system and used as media-related concepts (e.g., generating a media program).
- FIG. 6 is a flow chart illustrating an example method 600 .
- the method 600 can include accessing, by a first computing system, a program schedule for a media program.
- the first computing system is the VPS 202 and/or a component thereof (e.g., the sequencing system 314 ).
- the method 600 can include identifying, by the first computing system, a second computing system, wherein the identifying is based on the second computing system being associated with an event scheduled in the accessed program schedule.
- the second computing system is the end-user device 304 .
- the method 600 can include determining, by the first computing system, an estimated time period remaining until the event is scheduled to begin.
- the method 600 can include causing, by the first computing system, the identified second computing system to present, via a user interface of the second computing system, the determined estimated time period remaining until the event is scheduled to begin.
Abstract
Description
- This disclosure claims priority to (i) U.S. Provisional Patent Application No. 62/194,171, titled “Video Production System with Social Media Features,” filed on Jul. 17, 2015, and (ii) U.S. Provisional Patent Application No. 62/242,593, titled “Video Production System with Content-Related Features,” filed on Oct. 16, 2015, both of which are hereby incorporated by reference in their entirety.
- In this disclosure, unless otherwise specified and/or unless the particular context clearly dictates otherwise, the terms “a” or “an” mean at least one, and the term “the” means the at least one.
- In one aspect, an example method is disclosed. The method includes (i) accessing, by a first computing system, a program schedule for a media program; (ii) identifying, by the first computing system, a second computing system, wherein the identifying is based on the second computing system being associated with an event scheduled in the accessed program schedule; (iii) determining, by the first computing system, an estimated time period remaining until the event is scheduled to begin; and (iv) causing, by the first computing system, the identified second computing system to present, via a user interface of the second computing system, the determined estimated time period remaining until the event is scheduled to begin.
- In another aspect, an example non-transitory computer-readable medium is disclosed. The computer-readable medium has stored thereon program instructions that upon execution by a processor, cause performance of a first set of acts including (i) accessing, by a first computing system, a program schedule for a media program; (ii) identifying, by the first computing system, a second computing system, wherein the identifying is based on the second computing system being associated with an event scheduled in the accessed program schedule; (iii) determining, by the first computing system, an estimated time period remaining until the event is scheduled to begin; and (iv) causing, by the first computing system, the identified second computing system to present, via a user interface of the second computing system, the determined estimated time period remaining until the event is scheduled to begin.
- In another aspect, an example computing system is disclosed. The computing system is configured for performing a set of acts including (i) accessing, by a first computing system, a program schedule for a media program; (ii) identifying, by the first computing system, a second computing system, wherein the identifying is based on the second computing system being associated with an event scheduled in the accessed program schedule; (iii) determining, by the first computing system, an estimated time period remaining until the event is scheduled to begin; and (iv) causing, by the first computing system, the identified second computing system to present, via a user interface of the second computing system, the determined estimated time period remaining until the event is scheduled to begin.
-
FIG. 1 is a simplified block diagram of an example computing device. -
FIG. 2 is a simplified block diagram of an example video system. -
FIG. 3 is a simplified block diagram of an example video production system. -
FIG. 4A is a simplified diagram of an example frame of video content, without content overlaid thereon. -
FIG. 4B is a simplified diagram of an example frame of video content, with content overlaid thereon. -
FIG. 5 is a simplified block diagram of an example program schedule. -
FIG. 6 is a flow chart of an example method. - A video-production system (VPS) can generate video content that can serve as or be part of a video program (e.g., a news program). The VPS can then transmit the video content to a video-broadcast system (VBS), which in turn can transmit the video content to a first end-user device for presentation of the video content to an end-user.
- The VPS can include various components to facilitate generating video content. For example, the VPS can include a video source, a DVE system, a scheduling system, and a sequencing system. The video source can generate video content, and can transmit the video content to the DVE system. The DVE system can use the video content and a DVE template to execute a DVE, which can cause the DVE system to generate new video content that is a modified version of the received video content. For example, the generated video content can include the received video content with local weather content overlaid thereon.
- The scheduling system can create a program schedule, perhaps based on input received from a user (e.g., a producer or technical director) via a user interface. The sequencing system can process records in the program schedule, and based on the processed records, can control one or more components of the VPS, such as the video source and the DVE system, to facilitate generating video content.
- In one example, the VPS can also include a second end-user device, a content system, and a character generator. The second end-user device can transmit media content to the content system. For example second end-user device can capture video content (e.g., using a camera) and can transmit the captured video content to the content system.
- After the content system receives media content, the content system can then forward the received media content to the character generator. Alternatively, the content system can store the media content and later can retrieve and transmit the media content to the character generator.
- The character generator can receive the media content from the content system. The character generator can then use the received media content to generate video content that includes the received media content. Further, the character generator can transmit the generated video content to the DVE system.
- The DVE system can receive the video content and can execute a DVE, which causes the DVE system to generate video content that includes the received video content. The generated video content can serve as or be part of a video program. Thus, in this way, the VPS can integrate captured video content into a video program.
- As discussed above, the second end-user device can transmit video content to the VPS, which the VPS can integrate into a video program. In one example, the second end-user device can capture video content and, at or about the same time that the video content is being captured (i.e., without significant delay), can transmit the captured video content to the content system. This is can be referred to as a live transmission. As with other events in a video program, this type of event can be scheduled in the program schedule for the video program.
- In the case where the VPS is scheduled to, at some later time, receive a live transmission of video content provided by the second end-user device and live broadcast that video content to an audience, certain issues can arise. For example, although some users (e.g., those at the site of the VPS) can have access to the relevant program schedule for this event, the user of the second end-user device may not have access to the program schedule. As a result, the end-user can become frustrated by not knowing when the user's video content will become part of a live broadcast. In addition, the user can grow tired and become unprepared when the time comes and the live broadcast is supposed to begin.
- To help avoid these issues, it may be desirable to provide the user with an estimated time period remaining until the live broadcast (or another event) is scheduled to be begin. The VPS, the VBS, and/or components thereof can perform various acts and/or functions to help provide this feature. First, the sequencing system can access a program schedule for a video program. The sequencing system can then identify the second end-user device, based on the second end-user device being associated with a later event scheduled in the accessed program schedule. As noted above, the event can be an event where video content captured by the end-user device is live broadcast to an audience.
- In one example, the program schedule can include, as data associated with the later event, an identifier of the second end-user device and/or information such as an IP address of the second end-user device that can allow the sequencing system to communicate with the second end-user device. The sequencing system can then determine an estimated time period remaining until the later event is scheduled to begin. The sequencing system can do this in various ways. For example, for each event in a set of linear events between a current event and the later event, the sequencing system can determine a respective estimated duration of that event. Then, the sequencing system can determine that the estimated time period remaining until the later event is scheduled to begin is a sum of the determined durations.
- Also, in the case where one or more commercial breaks are scheduled to fall in between the current event and the later event, the sequencing system can determine the duration of each of the one or more commercial breaks and can add these one or more durations to the sum of the durations of the events as discussed above, to arrive at the estimated time period remaining until the later event is scheduled to begin.
- After the sequencing system determines the estimated time period remaining until the later event is scheduled to begin, the sequencing system can then cause the identified second end-user device to present the determined estimated time period. As such, in one example, the
sequencing system 314 can cause the end-user device 304 to display “ESTIMATED TIME REMAINING UNTIL YOU ARE ON THE AIR LIVE: 2 MINUTES, 30 SECONDS.” In some cases, the sequencing system can periodically or continuously perform one or more of the acts described in connection with this feature to allow the second end-user device to periodically or continuously display an updated estimated time period remaining until the later event is scheduled to begin. - These features and related features are described in greater detail below.
-
FIG. 1 is a simplified block diagram of anexample computing device 100. The computing device can be configured to perform and/or can perform one or more acts and/or functions, such as those described in this disclosure. Thecomputing device 100 can include various components, such as aprocessor 102, adata storage unit 104, acommunication interface 106, and/or auser interface 108. Each of these components can be connected to each other via aconnection mechanism 110. - In this disclosure, the term “connection mechanism” means a mechanism that facilitates communication between two or more components, devices, systems, or other entities. A connection mechanism can be a relatively simple mechanism, such as a cable or system bus, or a relatively complex mechanism, such as a packet-based communication network (e.g., the Internet). In some instances, a connection mechanism can include a non-tangible medium (e.g., in the case where the connection is wireless).
- The
processor 102 can include a general-purpose processor (e.g., a microprocessor) and/or a special-purpose processor (e.g., a digital signal processor (DSP)). Theprocessor 102 can execute program instructions contained in thedata storage unit 104 as discussed below. - The
data storage unit 104 can include one or more volatile, non-volatile, removable, and/or non-removable storage components, such as magnetic, optical, and/or flash storage, and/or can be integrated in whole or in part with theprocessor 102. Further, thedata storage unit 104 can take the form of a non-transitory computer-readable storage medium, having stored thereon program instructions (e.g., compiled or non-compiled program logic and/or machine code) that, upon execution by theprocessor 102, cause thecomputing device 100 to perform one or more acts and/or functions, such as those described in this disclosure. These program instructions can define and/or be part of a discrete software application. In some instances, thecomputing device 100 can execute program instructions in response to receiving an input, such as from thecommunication interface 106 and/or theuser interface 108. Thedata storage unit 104 can also store other types of data, such as those types described in this disclosure. - The
communication interface 106 can allow thecomputing device 100 to connect with and/or communicate with another other entity according to one or more protocols. In one example, thecommunication interface 106 can be a wired interface, such as an Ethernet interface or a high-definition serial-digital-interface (HD-SDI). In another example, thecommunication interface 106 can be a wireless interface, such as a cellular or WI-FI interface. In this disclosure, a connection can be a direct connection or an indirect connection, the latter being a connection that passes through and/or traverses one or more entities, such as a router, switcher, or other network device. Likewise, in this disclosure, a transmission can be a direct transmission or an indirect transmission. - The
user interface 108 can include hardware and/or software components that facilitate interaction between thecomputing device 100 and a user of thecomputing device 100, if applicable. As such, theuser interface 108 can include input components such as a keyboard, a keypad, a mouse, a touch-sensitive panel, and/or a media capturing device (e.g., a microphone and/or a camera), and/or output components such as a display device (which, for example, can be combined with a touch-sensitive panel), a sound speaker, and/or a haptic feedback system. - The
computing device 100 can take various forms, such as a workstation terminal, a desktop computer, a laptop, a tablet, a mobile phone, a set-top box, and/or a television. -
FIG. 2 is a simplified block diagram of anexample video system 200. Thevideo system 200 can perform various acts and/or functions related to video content, and can be implemented as a computing system. In this disclosure, the term “computing system” means a system that includes at least one computing device. In some instances, a computing system can include one or more other computing systems. - The
video system 200 can include various components, such as aVPS 202, aVBS 204, and an end-user device 206, each of which can be implemented as a computing system. Thevideo system 200 can also include aconnection mechanism 208, which connects theVPS 202 with theVBS 204; and aconnection mechanism 210, which connects theVBS 204 with the end-user device 206. -
FIG. 3 is a simplified block diagram of anexample VPS 202. TheVPS 202 can include various components, such as avideo source 302, an end-user device 304, acontent system 306, acharacter generator 308, a digital video-effect (DVE)system 310, ascheduling system 312, and asequencing system 314, each of which can be implemented as a computing system. TheVPS 202 can also include aconnection mechanism 316, which connects thevideo source 302 with thesequencing system 314; aconnection mechanism 318, which connects thevideo source 302 with theDVE system 310; aconnection mechanism 319, which connects the end-user device 304 with thesequencing system 314; aconnection mechanism 320, which connects the end-user device 304 with thecontent system 306;connection mechanism 322, which connects thecontent system 306 with thesequencing system 314; aconnection mechanism 324, which connects thecontent system 306 with thecharacter generator 308; aconnection mechanism 326, which connects thecharacter generator 308 with thesequencing system 314; aconnection mechanism 328, which connects thecharacter generator 308 with theDVE system 310; aconnection mechanism 330, which connects theDVE system 310 with thesequencing system 314; and aconnection mechanism 332, which connects thescheduling system 312 with thesequencing system 314. - The
video source 302 can take various forms, such as a video server, a video camera, a satellite receiver, a character generator, or a DVE system. An example video server is the K2 server provided by Grass Valley of San Francisco, Calif. - The
character generator 308 can take various forms. An example character generator is the VIZ TRIO provided by Viz Rt of Bergen, Norway. Another example character generator is CASPAR CG developed and distributed by the Swedish Broadcasting Corporation (SVT). - The
DVE system 310 can take various forms, such as a production switcher. An example production switcher is the VISION OCTANE production switcher provided by Ross Video Ltd. of Iroquois, Ontario in Canada. - The
scheduling system 312 can take various forms. An example scheduling system is WO TRAFFIC provided by WideOrbit, Inc. of San Francisco, Calif. Another example scheduling system is OSI-TRAFFIC provided by Harris Corporation of Melbourne, Fla. - The
sequencing system 314 can take various forms. A sequencing system is sometimes referred to in the industry as a “production automation system.” - Referring back to
FIG. 2 , theVBS 204 can include various components, such as a terrestrial antenna or a satellite transmitter, each of which can be implemented as a computing system. - Each of the video-based entities described in this disclosure can include or be integrated with a corresponding audio-based entity. Also, the video content described in this disclosure can include or be integrated with corresponding audio content. More generally, the entities can be media-based entities such as a media production system (MPS) and a media broadcast system (MBS).
- The
video system 200 and/or components thereof can perform various acts and/or functions. These features and related features will now be described. - The
video system 200 can perform various acts and/or functions related to video content. For example, thevideo system 200 can receive, generate, output, and/or transmit video content that can serve as or be part of a video program (e.g., a news program). In this disclosure, the act of receiving, generating, outputting, and/or transmitting video content can occur in various ways and/or according to various standards. For example, the act of receiving, outputting, and/or transmitting video content can include receiving, outputting, and/or transmitting a video stream representing the video content, such as over Internet Protocol (IP) or in accordance with the high-definition serial digital interface (HD-SDI) standard. Likewise, the act of generating content can include generating a video stream representing the video content. Also, the act of receiving, generating, outputting, and/or transmitting video content can include receiving, generating, outputting, and/or transmitting an encoded or decoded version of the video content. - The
VPS 202 can perform various acts and/or functions related to video content production. For example, theVPS 202 can generate and/or output video content, and can transmit the video content to another entity, such as theVBS 204. - Referring back to
FIG. 3 , within theVPS 202, thevideo source 302 can generate and/or output video content, and can transmit the video content to another entity, such as theDVE system 310. In practice, theVPS 202 is likely to include multiple video sources and corresponding connection mechanisms, each connecting a respective one of the video sources with theDVE system 310. - As noted above, the
video source 302 can take the form of a video server. A video server can record and/or store video content (e.g., in the form of a file). Further, the video server can retrieve stored video content and can use the retrieved video content to generate and/or output a video stream representing the video content. This is sometimes referred to in the industry as the video server playing out the video content. Thevideo server 302 can then transmit the video stream, thereby transmitting the video content, to another entity, such as theDVE system 310. - The end-
user device 304 can perform various acts and/or functions related to media content, perhaps based on input received from a user (e.g., a field reporter) via a user interface. For example, the end-user device 304 can obtain and/or generate media content. In one example, the end-user device can to this by using a video capturing device (e.g., a camera) of the end-user device 304 to capture video content. In one use case, this can allow the end-user device 304 to capture video content of a newsworthy event, such that it can be integrated into a news program. In another example, the end-user device 304 can download media content from a media source. - The end-
user device 304 can then transmit the media content to another entity, such as thecontent system 306. In one example, the end-user device 304 can store the media content in a data storage unit (e.g., a data storage unit of the end-user device 304). At a later time, the end-user device 304 can select and/or retrieve the stored media content, and can transmit it to another entity, such as thecontent system 306. In one example, the end-user device 304 can use one software application to capture and store video content, and can then use another software application to select, retrieve, and transmit the video content. - In another example, the end-
user device 304 can capture video content and, at or about the same time that the video content is being captured (i.e., without significant delay), can transmit the captured video content to another entity, such as thecontent system 306. This is sometimes referred to in the industry as a live video transmission or a live video stream. Notably, even when the end-user device 304 is transmitting video content in this manner, the end-user device 304 can still store, select, and/or retrieve the video content as part of this process. For instance, the end-user device 304 can store the video content in, and can retrieve it from, a memory buffer to facilitate the receipt and transmission of the video content. In one example, the end-user device 304 can use one software application to both capture and transmit video content in this manner. - In some cases, the end-
user device 304 can determine data related to captured video content, which the end-user device 304 can transmit along with the video content (e.g., as metadata). For example, the end-user device 304 can determine a location of the end-user device 304 (e.g., in the form of global positioning system (GPS) coordinates) where it was capturing video content. As another example, the end-user device 304 can determine a date and/or time when the end-user device 304 was capturing video content. - Like the end-
user device 304, thecontent system 306 can perform various acts and/or functions related to media content, perhaps based on input received from a user (e.g., a producer or technical director) via a user interface. For example, thecontent system 306 can receive media content and can do so in various ways. In one example, thecontent system 306 can receive media content from another entity, such as the end-user device 304. - The
content system 306 can also store, select, and/or retrieve media content. As such, thecontent system 306 can store received media content in a data storage unit (e.g., a data storage unit of the content system 306), and can then receive the media content by selecting and retrieving it from the data storage unit. Thecontent system 306 can also perform similar acts in connection with data relating to media content, such as data relating to video content, as discussed above. - The
content system 306 can also modify media content and/or related data, and can do so in various ways. In one example, thecontent system 306 can modify video content using a video-editing software application. Among other things, this can allow thecontent system 306 to removing vulgarities, personal information, and/or extraneous information that is not suitable or desirable for integration into a video program. In another example, the content system can modify the related data by adding a unique identification (ID) number to the data to facilitate managing the corresponding media content. In another example, the content system can add a tag, keyword, or description to the corresponding media content. - The
content system 306 can also transmit media content to another entity, such as thecharacter generator 308. In some cases, receiving and transmitting media content can include forwarding the media content. In other cases, receiving and transmitting media content can include receiving the media content and transmitting a copy of the media content. As such, in one example, thecontent system 306 can receive video content from the end-user device 304, and can transmit a copy of the video content to thecharacter generator 308. - In one example, the
content system 306 can receive media content from the end-user device 304 and, at or about the same time that the media content is being received (i.e., without significant delay), can transmit the media content to another entity, such as thecharacter generator 308. Notably, even when thecontent system 306 is transmitting media content in this manner, thecontent system 306 can still store, select, and/or retrieve the media content as part of this process. For instance, thecontent system 306 can store the media content in, and can retrieve it from, a memory buffer to facilitate the receipt and transmission of the media content. - The
content system 306 can also perform various acts and/or functions related to SM content. In this disclosure, “SM content” is content that has been published on a SM platform, which is a computer-based tool that allows users to create, share, and/or exchange content (e.g., in the form of text, images, and/or videos) in virtual communities on a computer-based network such as the Internet. Examples of SM platforms include TWITTER, YOUTUBE, FACEBOOK, PERISCOPE, INSTAGRAM, MEERKAT, LINKEDIN, and GOOGLE+. - The
content system 306 can receive SM content and can do so in various ways. For example, thecontent system 306 can receive SM content by obtaining it from another entity, such as a SM platform. In one example, thecontent system 306 can obtain SM content directly from a SM platform. In another example, thecontent system 306 can obtain SM content from a SM platform via a SM dashboard application (e.g., TWEETDECK, CYFE, or HOOTSUITE). In some instances, a SM dashboard application can provide additional searching and browsing functionalities (e.g., based on trend analysis or analytics) that may not be provided by the SM platform itself, and/or can provide access to multiple SM platforms through a single user interface. - SM content can include various elements such as (i) data indicating the SM platform from which the SM content was received, (ii) data identifying the publisher of the SM content (e.g., an account identifier, such as a username), (iii) a profile image corresponding to the publisher of the SM content, (iv) text published by the publisher in connection with the SM content, (v) an image published by the publisher in connection with the SM content, (vi) audio content published by the publisher in connection with the SM content, (vii) video content published by the publisher in connection with the SM content (viii) a timestamp indicating a time and/or date at which the SM content was published on the SM platform, (ix) a location (e.g., represented by global positioning system (GPS) coordinates) of the publisher when the SM content was published, (x) a location at which an aspect of the SM content occurred (e.g., where video content was recorded or where a photograph was taken), (xi) a timestamp indicating when an aspect of the SM content occurred, (xii) a number of other users associated with the publisher on a SM platform (e.g., a number of friends or followers), (xiii) an indication of how long the publisher has been a user of a SM platform, (xiv) a number of times the SM content has been shared (e.g., retweeted) by other users of a SM platform, (xv) a number of posts by the publisher on a SM platform, and/or (xvi) any other data that can be integrated into a video program.
- The SM system can also store, select, and/or retrieve SM content, perhaps based on input received from a user (e.g., a producer or technical director) via a user interface. As such, the
content system 306 can store obtained SM content in a data storage unit (e.g., a data storage unit of the content system 306), and can then receive the SM content by selecting and retrieving it from the data storage unit. - In some instances, the
content system 306 can select and modify SM content. Thecontent system 306 can select SM content in various ways. For example, thecontent system 306 can select SM content responsive to thecontent system 306 performing an action in connection with the SM content (e.g., responsive to thecontent system 306 receiving or storing the SM content). In another example, thecontent system 306 can select SM content based on the SM content being associated with a particular characteristic (e.g., based on the SM content being scheduled to be integrated into a video program). In another example, thecontent system 306 can, periodically or based on a schedule, select SM content for routine processing. As yet another example, thecontent system 306 can select SM content based on input received from a user via a user interface. - The
content system 306 can then modify the selected SM content by identifying a first element of the selected SM content based on the first element being associated with a particular characteristic, and then modifying the selected SM content by modifying the identified first element of the selected SM content item. - The
character generator 308 can perform various acts and/or functions, perhaps based on input received via a user interface. For example, thecharacter generator 308 can receive media content and can use a character generator template and received media content to generate and/or output video content that includes the received media content. Thecontent system 306 can receive media content in various ways. In one example, thecontent system 306 can receive media content from another entity. For example, thecontent system 306 can receive video content (e.g., video content originally captured by the end-user device 304) from thecontent system 306. In another example, thecharacter generator 308 can receive SM content from thecontent system 306. In another example, thecharacter generator 308 can receive media content by selecting and retrieving it from a data storage unit (e.g., a data storage unit of the content system 306). - The character generator template specifies the manner in which the
character generator 308 uses the received media content to generate and/or output the video content. Thecharacter generator 308 can create and/or modify a character generator template. Further, thecharacter generator 308 can store, select, and/or retrieve a character generator template. As such, thecharacter generator 308 can store a character generator template in a data storage unit (e.g., a data storage unit of the character generator 308), and can then receive the character generator template by retrieving it from the data storage unit. - The character generator template can specify how the
character generator 308 is to receive media content. In one example, the character generator template can do so by specifying that thecharacter generator 308 is to receive media content on a particular input of the character generator 308 (e.g., an input that maps to a particular entity, such as the content system 306). In another example, the character generator template can do so by specifying that thecharacter generator 308 is to receive media content by retrieving it from a particular location of a particular data storage unit (e.g., a data storage unit of the character generator 308). - In one example, the
character generator 308 can use an ordered set of media content items to generate video content that includes the media content items in the specified order. This type of generated video content is sometimes referred to in the industry as a “ticker.” The media content items can include various types of content, such as text and/or images. The ordered set of media content items can be stored in various forms, such as in the form of an Extensible Markup Language (XML) file. - After the
character generator 308 generates and/or outputs video content, thecharacter generator 308 can transmit the video content to another entity, such as theDVE system 310, and/or can store the video content in a data storage unit (e.g., a data storage unit of the character generator 308). - As such, in one example, the
character generator 308 can receive media content, can use the media content to generate and/or output video content that includes the media content, and can transmit the video content to theDVE system 310. - The
DVE system 310 can use a DVE template to generate and/or output video content. This is sometimes referred to in the industry as the DVE system “executing a DVE.” In some instances, theDVE system 310 can execute multiple DVEs in serial or overlapping fashion. - The DVE template specifies the manner in which the
DVE system 310 generates and/or outputs video content. TheDVE system 310 can create and/or modify a DVE template, perhaps based on input received from a user via a user interface. Further, theDVE system 310 can store and/or retrieve a DVE template, perhaps based on input received from a user via a user interface. As such, theDVE system 310 can store a DVE system template in a data storage unit (e.g., a data storage unit of the DVE system 310), and can then receive the DVE template by selecting and retrieving it from the data storage unit. - In some instances, the
DVE system 310 can use the DVE template and media content to generate and/or output video content that includes the media content. TheDVE system 310 can receive content in various ways. For example, theDVE system 310 can do so by receiving it from another entity, such as thevideo source 302 and/or thecharacter generator 308. In another example, theDVE system 310 can do so by selecting and retrieving it from a data storage unit (e.g., a data storage unit of the DVE system 310). - The DVE template can specify how the
DVE system 310 is to receive media content. In one example, the DVE template can do so by specifying that theDVE system 310 is to receive media content on a particular input of the DVE system 310 (e.g., an input that maps to a particular entity, such as thevideo source 302 or the character generator 308). In another example, the DVE template can do so by specifying that theDVE system 310 is to receive content by retrieving it from a particular location of a particular data storage unit (e.g., a data storage unit of the DVE system 310). - A DVE template can be configured in various ways, which can allow the
DVE system 310 to execute various types of DVEs. In one example, a DVE template can specify that theDVE system 310 is to receive video content from thevideo source 302 and other media content (e.g., local weather content) from a data storage unit of the DVE system, and is to overlay the other media content on the video content, thereby generating a modified version of the video content. As such, in one example, theDVE system 310 can generate video content by modifying video content. -
FIGS. 4A and 4B help illustrate this concept of overlaying other content on video content.FIG. 4A is a simplified depiction of anexample frame 400 of video content.Frame 400 includescontent 402, but does not include other content overlaid oncontent 402. For comparison,FIG. 4B is a simplified depiction of anotherexample frame 450 of video content.Frame 450 includescontent 452 andother content 454 overlaid oncontent 452. - In another example, a DVE template can specify that the
DVE system 310 is to receive first video content from thevideo source 302 and second video content from thecharacter generator 308, and is to overlay the second video content on the first video content, thereby generating a modified version of the first video content. - In another example, a DVE template can specify that the
DVE system 310 is to receive first video content from thevideo source 302 and second video content from thecharacter generator 308, and is to scale-down and re-position the first video content and the second video content, each in a respective one of two windows positioned side-by-side. As such, theDVE system 310 can generate video content by scaling and/or re-positioning video content. - After the
DVE system 310 generates and/or outputs the video content, theDVE system 310 can transmit the video content to another entity, such as theVBS 204, or can store the video content in a data storage unit (e.g., a data storage unit of the DVE system 310). - As such, in one example, the
DVE system 310 can receive first video content including video content, can use the first video content to generate and/or output second video content that includes the video content. This is an example way in which theVPS 202 can integrate video content into a video program. - The
VPS 202 can also integrate video content into a video program in other ways. For example, in the case where thevideo source 302 is a video camera, thecontent system 306 can include a display device that is located within a field of the view of the video camera while the video camera records video content that serves as or is made part of the video program. In one example, the display device can be touch-enabled, which can allow a user (e.g., a news anchor) to interact with the video content. To facilitate the user's interaction with the video content, the display device and/or other components of thecontent system 306 can be programmed with instructions that cause particular actions in response to particular touch commands. - The
scheduling system 312 can perform various acts and/or functions related to the scheduling of video content production. For example, thescheduling system 312 can create and/or modify a program schedule of a video program, perhaps based on input received from a user via a user interface. Further, thescheduling system 312 can store and/or retrieve a program schedule, perhaps based on input received from a user via a user interface. As such, thescheduling system 312 can store a program schedule in a data storage unit (e.g., a data storage unit of the scheduling system 312), and can then receive the program schedule by selecting and retrieving it from the data storage unit. Thescheduling system 312 can also transmit a program schedule to another entity, such as thesequencing system 314. - The
sequencing system 314 can process records in the program schedule. This can cause thesequencing system 314 to control one or more other components of theVPS 202 to facilitate theVPS 202 generating and/or outputting video content, which can serve as or be part of a video program. For example, thesequencing system 314 can control thevideo source 302, thecontent system 306, thecharacter generator 308, and/or theDVE system 310 to perform the various acts and/or functions described in this disclosure. - The
sequencing system 314 can receive a program schedule in various ways. For example, thesequencing system 314 can do so by receiving it from another entity, such as thescheduling system 312. In another example, thecharacter generator 308 can do so by selecting and retrieving it from a data storage unit (e.g., a data storage unit of the scheduling system 312). - A program schedule (sometimes referred to in the industry as a “rundown”) serves as a schedule or outline of a video program. A program schedule can thus schedule one or more events associated with the production of a video program. A program schedule can include multiple records. A video program can be conceptually divided into multiple logically-separated portions (sometimes referred to in the industry as “stories”). As such, each portion of the video program can be represented by a separate record of the program schedule. In some cases, each record can also include one or more sub-records. Each record (including a sub-record) can include various types of data.
-
FIG. 5 is a simplified diagram of an example program schedule 500. The program schedule 500 includes ten records represented as ten ordered rows. Each record corresponds to a respective portion of a video program, except for one which corresponds to a commercial break. For each portion, the respective record specifies at least one data item that corresponds to that portion of the video program. In particular, each record specifies at least one of a story title, a video content item identifier, a duration, and a script. - In some cases, the duration can be an estimated duration, as discussed below. The script can include various instructions, such as (i) instructions for the
video source 302 to playout video content, (ii) instructions for thecharacter generator 308 to generate video content based on stored video content and/or video content received bycontent system 306 as a live transmission from the end-user device 304, and forwarded to thecharacter generator 308, and (iii) instructions for theDVE system 310 to execute a particular DVE that overlays video content on other video content. The script can also include other information such as text scheduled to be displayed by a teleprompter and to be real aloud by a news anchor. - A video content item can consist of logically-related video content. For example, a video content item can be a commercial. As another example, a video content item can be a portion of a television program that is scheduled between two commercial breaks. This is sometimes referred to in the industry as a “program segment.”
- As shown in
FIG. 5 , the first record specifies a story title of STORY A, a video content identifier of VCI ID A, a duration of 00:02:00:00 (in hours::minutes::seconds::frames format), and a script of SCRIPT A. In this example, SCRIPT A includes instructions for thevideo source 302 to playout a first video content item identified by the identifier VCI A for two minutes, instructions for thecharacter generator 308 to use a live transmission of a second video content item provided by the end-user device 304 to generate a third video content item that includes the second video content item, and instructions for theDVE system 310 to execute a particular DVE, which causes theDVE system 310 to overlay the generated third video content item on the generated first video content item to generate a fourth video content item. - The program schedule 500 has been greatly simplified for the purposes of illustrating certain features. In practice, a program schedule is likely to include significantly more data.
- In some instances, the
sequencing system 314 can process a next record (or a portion thereof) in the program schedule based on a trigger event. In one example, the trigger event can be thesequencing system 314 completing one or more actions related to a current record in the program schedule. In another example, the trigger event can be thesequencing system 314 receiving input from a user via a user interface. - Referring back to
FIG. 2 , theVBS 204 can receive video content from theVPS 202, which in turn can transmit the video content to the end-user device 206 for presentation of the video content to an end user. In practice, theVBS 204 can transmit video content to a large number of end-user devices for presentation of the video content to a large number of end users. TheVBS 204 can transmit video content to the end-user device 206 in various ways. For example,VBS 204 can transmit video content to the end-user device 206 over-the-air or via a packet-based network such as the Internet. The end-user device 206 can receive video content from theVBS 204, and can present the video content to an end user via a user interface. - As noted above, the end-
user device 304 can capture video content and, at or about the same time that the video content is being captured (i.e., without significant delay), can transmit the captured video content to another entity, such as thecontent system 306. In some cases, some or all of the other operations described in this disclose can also be performed at or about that same time, which can allow theVBS 204 to transmit the captured video content to the end-user device 206 as a live transmission. - In the case where the
VPS 202 is scheduled to, at some later time, receive a live transmission of video content provided by the end-user device 304 and live broadcast that video content to an audience, certain issues can arise. For example, although some users (e.g., those at the site of the VPS 202) can have access to the relevant program schedule for this event, the user of the end-user device 304 may not have access to the program schedule. As a result, the user can become frustrated by not knowing when the user's video content will become part of a live broadcast. In addition, the user can grow tired and become unprepared when the time comes and the live broadcast is supposed to begin. - To help avoid these issues, it may be desirable to provide the user with an estimated time period remaining until the live broadcast (or another event) is scheduled to begin. The
VPS 202 and/or theVBS 204 can perform various acts and/or functions to help provide this feature. These acts and/or functions and related features will now be described. - First, the
sequencing system 314 can access a program schedule for a video program. Thesequencing system 314 can then identify the end-user device 304, based on the end-user device 304 being associated with a later event scheduled in the accessed program schedule. As noted above, the event can be an event where video content captured by the end-user device 304 is live broadcast to an audience. - In one example, the program schedule can include, as data associated with the later event, an identifier of the end-
user device 304 and/or information such as an IP address of the end-user device 304 that can allow thesequencing system 314 to communicate with the end-user device 304. Additionally or alternatively, thesequencing system 314 can include, as data associated with the later event, an identifier of a video stream (representing video content) that the end-user device 304 is already live transmitting to theVPS 202. In this case, thesequencing system 314 can use mapping data (which thecontent system 306 can maintain for all active video streams) to map the identifier of the video stream to an identifier of the end-user device 304. As such, in one example, thesequencing system 314 can identify a video stream associated with the later event, and can use the identified video stream as a basis to identify the end-user device 304. - The
sequencing system 314 can then determine an estimated time period remaining until the later event is scheduled to begin. Thesequencing system 314 can do this in various ways. For example, for each event in a set of linear events between a current event and the later event, thesequencing system 314 can determine a respective estimated duration of that event. Then, thesequencing system 314 can determine that the estimated time period remaining until the later event is scheduled to begin is a sum of the determined durations. In some instances, thesequencing system 314 can also add an estimated time period remaining in the current event to this sum. - The
sequencing system 314 can estimate the duration of a given event in various ways. For example, in the case where an event has a predefined duration (e.g., where the event involves playout of a video content item having a duration of 20 seconds), thesequencing system 314 can use the predefined duration as the estimated duration. Alternatively, in the case where the duration of an event has an unknown or variable duration, thesequencing system 314 can estimate the duration of the event based on data related to the event. In some cases, thesequencing system 314 can obtain this data from the program schedule. For instance, in the case where an event involves a news anchor reading text from a teleprompter, thesequencing system 314 can use the number of words in the text to estimate how long it will take anchor to read the text. - Also, in the case where one or more commercial breaks are scheduled to fall in between the current event and the later event, the
sequencing system 314 can determine the duration of each of the one or more commercial breaks and can add these one or more durations to the sum of the durations of the events as discussed above, to arrive at the estimated time period remaining until the later event is scheduled to begin. In some cases, thesequencing system 314 can determine the number and/or duration of any applicable commercial breaks based on data specified in the program schedule and/or in a corresponding broadcast automation schedule (sometimes referred to in the industry as a “traffic log”). - After the
sequencing system 314 determines the estimated time period remaining until the later event is scheduled to begin, thesequencing system 314 can then cause the identified end-user device 304 to present, via a user interface of the end-user device 304, the determined estimated time period. Thesequencing system 314 can cause the identified end-user device 304 to present the determined estimated time period in various ways. For example, thesequencing system 314 can cause this to happen by transmitting, to the end-user device 304, an instruction that causes the end-user device 304 to present the determined estimated time period. In one example, the instruction can include the determined estimated time period or other data that can allow the end-user device 304 to determine the same. - The end-
user device 304 can present the determined estimated time period in various ways, such as by displaying it on a display screen of the end-user device 304 or by providing it audibly through a speaker of the end-user device 304. As such, in one example, thesequencing system 314 can cause the end-user device 304 to display “ESTIMATED TIME REMAINING UNTIL YOU ARE ON THE AIR LIVE: 2 MINUTES, 30 SECONDS.” In some cases, thesequencing system 314 can periodically or continuously perform one or more of the acts described in connection with this feature to allow the end-user device 304 to periodically or continuously display an updated estimated time period remaining until the later event is scheduled to begin. - Additionally or alternatively, after the (now) current event begins, the
sequencing system 314 can determine an estimated time period remaining until the current event is scheduled to end. Thesequencing system 314 can do this in various ways. For example, where the program schedule specifies a predefined or target duration, thesequencing system 314 can use this duration and a time period that has passed since the current event began to determine an estimated time period remaining until the current event is scheduled to end. - The
sequencing system 314 can then cause the identified end-user device 304 to present, via a user interface of the end-user device 304, the estimated time period remaining until the current event is scheduled to end. Thesequencing system 314 can cause the identified end-user device 304 to present this determined estimated time period in various ways. For example, thesequencing system 314 can cause this to happen by transmitting, to the end-user device 304, an instruction that causes the end-user device 304 to present the determined estimated time period. In one example, the instruction can include the determined estimated time period or other data that can allow the end-user device 304 to determine the same. - As with the other determined estimated time, the end-
user device 304 can present this determined estimated time period in various ways, such as by displaying it on a display screen of the end-user device 304 or by providing it audibly through a speaker of the end-user device 304. As such, in one example, thesequencing system 314 can cause the end-user device 304 to display “ESTIMATED TIME REMAINING UNTIL YOU ARE OFF THE AIR: 20 SECONDS.” In some cases, thesequencing system 314 can periodically or continuously perform one or more of the acts described in connection with this feature to allow the end-user device 304 to periodically or continuously display an updated estimated time period remaining until the current event is scheduled to end. - Although some of the acts and/or functions described in this disclosure have been described in the context of the
video system 200 and as video-related concepts (e.g., generating a video program), the acts and/or functions can also be applied in the context of an audio system and used as audio-related concepts (e.g., generating an audio program). More generally, the acts and/or functions described in this disclosure can be applied in the context of a media system and used as media-related concepts (e.g., generating a media program). -
FIG. 6 is a flow chart illustrating anexample method 600. - At
block 602, themethod 600 can include accessing, by a first computing system, a program schedule for a media program. In one example, the first computing system is theVPS 202 and/or a component thereof (e.g., the sequencing system 314). - At
block 604, themethod 600 can include identifying, by the first computing system, a second computing system, wherein the identifying is based on the second computing system being associated with an event scheduled in the accessed program schedule. In one example, the second computing system is the end-user device 304. - At
block 606, themethod 600 can include determining, by the first computing system, an estimated time period remaining until the event is scheduled to begin. - At
block 608, themethod 600 can include causing, by the first computing system, the identified second computing system to present, via a user interface of the second computing system, the determined estimated time period remaining until the event is scheduled to begin. - Although some of the acts and/or functions described in this disclosure have been described as being performed by a particular entity, the acts and/or functions can be performed by any entity, such as those entities described in this disclosure. Further, although the acts and/or functions have been recited in a particular order, the acts and/or functions need not be performed in the order recited. However, in some instances, it can be desired to perform the acts and/or functions in the order recited. Further, each of the acts and/or functions can be performed responsive to one or more of the other acts and/or functions. Also, not all of the acts and/or functions need to be performed to achieve one or more of the benefits provided by this disclosure, and therefore not all of the acts and/or functions are required.
- Although certain variations have been discussed in connection with one or more example of this disclosure, these variations can also be applied to all of the other examples of this disclosure as well.
- Although select examples of this disclosure have been described, alterations and permutations of these examples will be apparent to those of ordinary skill in the art. Other changes, substitutions, and/or alterations are also possible without departing from the invention in its broader aspects as set forth in the following claims.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/213,126 US20170019715A1 (en) | 2015-07-17 | 2016-07-18 | Media production system with scheduling feature |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201562194171P | 2015-07-17 | 2015-07-17 | |
US201562242593P | 2015-10-16 | 2015-10-16 | |
US15/213,126 US20170019715A1 (en) | 2015-07-17 | 2016-07-18 | Media production system with scheduling feature |
Publications (1)
Publication Number | Publication Date |
---|---|
US20170019715A1 true US20170019715A1 (en) | 2017-01-19 |
Family
ID=57775229
Family Applications (18)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/211,153 Expired - Fee Related US10291679B2 (en) | 2015-07-17 | 2016-07-15 | Permission request for social media content in a video production system |
US15/211,252 Expired - Fee Related US10412142B2 (en) | 2015-07-17 | 2016-07-15 | Video-production system with social-media features |
US15/211,170 Abandoned US20170019450A1 (en) | 2015-07-17 | 2016-07-15 | Media production system with social media feature |
US15/211,243 Abandoned US20170017371A1 (en) | 2015-07-17 | 2016-07-15 | Video-Production System With Social-Media Features |
US15/211,165 Abandoned US20170019694A1 (en) | 2015-07-17 | 2016-07-15 | Video production system with social media content item modification feature |
US15/211,149 Expired - Fee Related US9979770B2 (en) | 2015-07-17 | 2016-07-15 | Method of requesting a social media link authorization based on a social media content item |
US15/211,155 Abandoned US20170019448A1 (en) | 2015-07-17 | 2016-07-15 | Media production system with social media content integration feature |
US15/211,189 Expired - Fee Related US10320876B2 (en) | 2015-07-17 | 2016-07-15 | Media production system with location-based feature |
US15/211,193 Expired - Fee Related US9961123B2 (en) | 2015-07-17 | 2016-07-15 | Media production system with score-based display feature |
US15/212,634 Expired - Fee Related US10305957B2 (en) | 2015-07-17 | 2016-07-18 | Video production system with DVE feature |
US15/213,126 Abandoned US20170019715A1 (en) | 2015-07-17 | 2016-07-18 | Media production system with scheduling feature |
US15/213,123 Abandoned US20170019697A1 (en) | 2015-07-17 | 2016-07-18 | Media production system with scheduling feature |
US15/927,904 Expired - Fee Related US10348794B2 (en) | 2015-07-17 | 2018-03-21 | Media production system with score-based display feature |
US16/366,527 Abandoned US20190222624A1 (en) | 2015-07-17 | 2019-03-27 | Permission Request For Social Media Content In A Video Production System |
US16/386,242 Expired - Fee Related US10581947B2 (en) | 2015-07-17 | 2019-04-16 | Video production system with DVE feature |
US16/397,739 Abandoned US20190253474A1 (en) | 2015-07-17 | 2019-04-29 | Media production system with location-based feature |
US16/452,022 Abandoned US20190320003A1 (en) | 2015-07-17 | 2019-06-25 | Media production system with score-based display feature |
US16/531,978 Abandoned US20190364085A1 (en) | 2015-07-17 | 2019-08-05 | Video-Production System With Social-Media Features |
Family Applications Before (10)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/211,153 Expired - Fee Related US10291679B2 (en) | 2015-07-17 | 2016-07-15 | Permission request for social media content in a video production system |
US15/211,252 Expired - Fee Related US10412142B2 (en) | 2015-07-17 | 2016-07-15 | Video-production system with social-media features |
US15/211,170 Abandoned US20170019450A1 (en) | 2015-07-17 | 2016-07-15 | Media production system with social media feature |
US15/211,243 Abandoned US20170017371A1 (en) | 2015-07-17 | 2016-07-15 | Video-Production System With Social-Media Features |
US15/211,165 Abandoned US20170019694A1 (en) | 2015-07-17 | 2016-07-15 | Video production system with social media content item modification feature |
US15/211,149 Expired - Fee Related US9979770B2 (en) | 2015-07-17 | 2016-07-15 | Method of requesting a social media link authorization based on a social media content item |
US15/211,155 Abandoned US20170019448A1 (en) | 2015-07-17 | 2016-07-15 | Media production system with social media content integration feature |
US15/211,189 Expired - Fee Related US10320876B2 (en) | 2015-07-17 | 2016-07-15 | Media production system with location-based feature |
US15/211,193 Expired - Fee Related US9961123B2 (en) | 2015-07-17 | 2016-07-15 | Media production system with score-based display feature |
US15/212,634 Expired - Fee Related US10305957B2 (en) | 2015-07-17 | 2016-07-18 | Video production system with DVE feature |
Family Applications After (7)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/213,123 Abandoned US20170019697A1 (en) | 2015-07-17 | 2016-07-18 | Media production system with scheduling feature |
US15/927,904 Expired - Fee Related US10348794B2 (en) | 2015-07-17 | 2018-03-21 | Media production system with score-based display feature |
US16/366,527 Abandoned US20190222624A1 (en) | 2015-07-17 | 2019-03-27 | Permission Request For Social Media Content In A Video Production System |
US16/386,242 Expired - Fee Related US10581947B2 (en) | 2015-07-17 | 2019-04-16 | Video production system with DVE feature |
US16/397,739 Abandoned US20190253474A1 (en) | 2015-07-17 | 2019-04-29 | Media production system with location-based feature |
US16/452,022 Abandoned US20190320003A1 (en) | 2015-07-17 | 2019-06-25 | Media production system with score-based display feature |
US16/531,978 Abandoned US20190364085A1 (en) | 2015-07-17 | 2019-08-05 | Video-Production System With Social-Media Features |
Country Status (5)
Country | Link |
---|---|
US (18) | US10291679B2 (en) |
EP (4) | EP3326377A4 (en) |
CA (5) | CA2992484A1 (en) |
IL (4) | IL256782A (en) |
WO (9) | WO2017015098A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10021442B1 (en) * | 2013-03-14 | 2018-07-10 | Tribune Broadcasting Company, Llc | Systems and methods for causing a stunt switcher to run a bug-removal DVE |
US10104449B1 (en) | 2013-03-14 | 2018-10-16 | Tribune Broadcasting Company, Llc | Systems and methods for causing a stunt switcher to run a bug-overlay DVE |
US10341737B2 (en) * | 2016-09-23 | 2019-07-02 | DISH Technologies L.L.C. | Integrating broadcast media streams with user media streams |
US20220321969A1 (en) * | 2021-04-02 | 2022-10-06 | Shanghai Bilibili Technology Co., Ltd. | Bullet comment processing method and apparatus |
Families Citing this family (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017015098A1 (en) * | 2015-07-17 | 2017-01-26 | Tribune Broadcasting Company, Llc | Video-production system with social-media features |
US9966110B2 (en) * | 2015-10-16 | 2018-05-08 | Tribune Broadcasting Company, Llc | Video-production system with DVE feature |
US10320728B2 (en) * | 2016-12-13 | 2019-06-11 | Google Llc | Methods, systems, and media for generating a notification in connection with a video content item |
CN106919707B (en) * | 2017-03-10 | 2020-04-07 | 腾讯科技(深圳)有限公司 | Page display method and terminal based on H5 |
US10648690B2 (en) * | 2017-04-14 | 2020-05-12 | Johnson Controls Technology Company | Multi-function thermostat with event schedule controls |
US10467710B2 (en) * | 2017-05-19 | 2019-11-05 | BlackBook Media Inc. | Social media platform enabling multiple social media aliases |
US10735808B2 (en) * | 2017-08-10 | 2020-08-04 | The Nielsen Company (Us), Llc | Methods and apparatus of media device detection for minimally invasive media meters |
CN109522434B (en) * | 2018-10-24 | 2021-08-03 | 武汉大学 | Social image geographic positioning method and system based on deep learning image retrieval |
US11449664B1 (en) * | 2019-07-01 | 2022-09-20 | Instasize, Inc. | Template for creating content item |
US11676316B1 (en) | 2019-07-01 | 2023-06-13 | Instasize, Inc. | Shareable settings for modifying images |
US10992602B2 (en) * | 2019-08-19 | 2021-04-27 | Landis+Gyr Innovations, Inc. | Sequential storage of collected data from heterogeneous intervals |
US20220408126A1 (en) * | 2021-06-18 | 2022-12-22 | Grass Valley Limited | System and method for optimizing the distribution of available media production resources |
WO2022263665A1 (en) * | 2021-06-18 | 2022-12-22 | Grass Valley Limited | System and method for optimizing the distribution of available media production resources |
US20220414325A1 (en) * | 2021-06-29 | 2022-12-29 | Klaviyo Inc. | Universal content for email template |
US20230136265A1 (en) * | 2021-10-29 | 2023-05-04 | International Business Machines Corporation | Content management system |
CN114418555B (en) * | 2022-03-28 | 2022-06-07 | 四川高速公路建设开发集团有限公司 | Project information management method and system applied to intelligent construction |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030122863A1 (en) * | 2001-12-28 | 2003-07-03 | International Business Machines Corporation | Navigation tool for slide presentations |
US20030149975A1 (en) * | 2002-02-05 | 2003-08-07 | Charles Eldering | Targeted advertising in on demand programming |
US20060156219A1 (en) * | 2001-06-27 | 2006-07-13 | Mci, Llc. | Method and system for providing distributed editing and storage of digital media over a network |
US20070106419A1 (en) * | 2005-09-07 | 2007-05-10 | Verizon Business Network Services Inc. | Method and system for video monitoring |
US20070168543A1 (en) * | 2004-06-07 | 2007-07-19 | Jason Krikorian | Capturing and Sharing Media Content |
US7733808B2 (en) * | 2006-11-10 | 2010-06-08 | Microsoft Corporation | Peer-to-peer aided live video sharing system |
US20120192225A1 (en) * | 2011-01-25 | 2012-07-26 | Youtoo Technologies, LLC | Administration of Content Creation and Distribution System |
US20130104177A1 (en) * | 2011-10-19 | 2013-04-25 | Google Inc. | Distributed real-time video processing |
US8539542B1 (en) * | 2009-08-25 | 2013-09-17 | Whdc Llc | System and method for managing multiple live video broadcasts via a public data network on a single viewing channel |
US20140341527A1 (en) * | 2013-05-15 | 2014-11-20 | MixBit, Inc. | Creating, Editing, and Publishing a Video Using a Mobile Device |
US20150215665A1 (en) * | 2014-01-30 | 2015-07-30 | Echostar Technologies L.L.C. | Methods and apparatus to synchronize second screen content with audio/video programming using closed captioning data |
US20150326814A1 (en) * | 2014-05-12 | 2015-11-12 | Echostar Uk Holdings Limited | Systems and method for timing commercial breaks |
US20160073029A1 (en) * | 2014-09-07 | 2016-03-10 | Guy MARKOVITZ | Method and system for creating a video |
US20160182979A1 (en) * | 2014-12-22 | 2016-06-23 | Sony Corporation | Method and system for processing closed-caption information |
US20160300594A1 (en) * | 2015-04-10 | 2016-10-13 | OMiro IP LLC | Video creation, editing, and sharing for social media |
Family Cites Families (155)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6868551B1 (en) * | 1995-05-05 | 2005-03-15 | Microsoft Corporation | Interactive program summary panel |
US5801787A (en) * | 1996-06-14 | 1998-09-01 | Starsight Telecast, Inc. | Television schedule system and method of operation for multiple program occurrences |
US6452612B1 (en) * | 1998-12-18 | 2002-09-17 | Parkervision, Inc. | Real time video production system and method |
US7835920B2 (en) * | 1998-12-18 | 2010-11-16 | Thomson Licensing | Director interface for production automation control |
US20030001880A1 (en) * | 2001-04-18 | 2003-01-02 | Parkervision, Inc. | Method, system, and computer program product for producing and distributing enhanced media |
US20040194131A1 (en) * | 1999-03-11 | 2004-09-30 | Ellis Michael D. | Television system with scheduling of advertisements |
AU2001249669A1 (en) * | 2000-03-31 | 2001-10-15 | Parkervision, Inc. | Method, system and computer program product for full news integration and automation in a real time video production environment |
US20020087987A1 (en) * | 2000-11-16 | 2002-07-04 | Dudkiewicz Gil Gavriel | System and method for creating and editing a viewer profile used in determining the desirability of video programming events |
US7168085B2 (en) * | 2002-01-31 | 2007-01-23 | Microsoft Corporation | Time-based selection of EPG data destined for low resource clients |
US20030237093A1 (en) * | 2002-06-19 | 2003-12-25 | Marsh David J. | Electronic program guide systems and methods for handling multiple users |
KR100567541B1 (en) * | 2003-07-08 | 2006-04-05 | 주식회사 팬택앤큐리텔 | Ststem and Method for management based on mobile network of the goods information broadcasted on home shopping channels |
CN100551033C (en) | 2003-11-12 | 2009-10-14 | 皇家飞利浦电子股份有限公司 | Program recommendation system |
JP4766920B2 (en) * | 2005-05-18 | 2011-09-07 | シャープ株式会社 | Broadcast program receiver |
US20060277457A1 (en) * | 2005-06-07 | 2006-12-07 | Salkind Carole T | Method and apparatus for integrating video into web logging |
EP1750439A1 (en) * | 2005-08-03 | 2007-02-07 | Thomson Licensing | EPG content analyser |
US9098545B2 (en) * | 2007-07-10 | 2015-08-04 | Raj Abhyanker | Hot news neighborhood banter in a geo-spatial social network |
US20080020814A1 (en) * | 2006-06-02 | 2008-01-24 | Nicolas Kernene | Broadband gaming system and method |
US7779444B2 (en) * | 2006-07-23 | 2010-08-17 | William Glad | System and method for video on request |
US20080077568A1 (en) * | 2006-09-26 | 2008-03-27 | Yahoo! Inc. | Talent identification system and method |
US20100293048A1 (en) * | 2006-10-19 | 2010-11-18 | Taboola.Com Ltd. | Method and system for content composition |
US20140136328A1 (en) * | 2006-11-22 | 2014-05-15 | Raj Abhyanker | Immediate communication between neighboring users surrounding a specific geographic location |
US20080134298A1 (en) | 2006-12-04 | 2008-06-05 | Microsoft Corporation | Integrated development environment with community integration |
WO2008121967A2 (en) * | 2007-03-30 | 2008-10-09 | Google Inc. | Interactive media display across devices |
EP2135156A4 (en) * | 2007-04-12 | 2011-02-02 | Thomson Licensing | Worklow engine for media production and distribution |
US7930420B2 (en) * | 2007-06-25 | 2011-04-19 | University Of Southern California | Source-based alert when streaming media of live event on computer network is of current interest and related feedback |
US20090006211A1 (en) * | 2007-07-01 | 2009-01-01 | Decisionmark Corp. | Network Content And Advertisement Distribution System and Method |
US20090037263A1 (en) * | 2007-07-30 | 2009-02-05 | Yahoo! Inc. | System for the insertion and control of advertisements in video |
US20090044235A1 (en) * | 2007-08-07 | 2009-02-12 | Davidson Daniel L | Method and system for on-line content acquisition and distribution |
US20110276423A1 (en) * | 2007-08-07 | 2011-11-10 | Onenews Corporation | Systems and Methods for Content Communication |
US8275764B2 (en) | 2007-08-24 | 2012-09-25 | Google Inc. | Recommending media programs based on media program popularity |
US20090064246A1 (en) * | 2007-08-30 | 2009-03-05 | Bell Gregory P | Distributed and interactive globecasting system |
US7970418B2 (en) * | 2007-08-31 | 2011-06-28 | Verizon Patent And Licensing Inc. | Method and system of providing event content sharing by mobile communication devices |
US8849183B2 (en) * | 2007-10-05 | 2014-09-30 | Qualcomm Incorporated | Location and time based filtering of broadcast information |
US20090100469A1 (en) * | 2007-10-15 | 2009-04-16 | Microsoft Corporation | Recommendations from Social Networks |
KR101404747B1 (en) * | 2007-11-16 | 2014-06-12 | 엘지전자 주식회사 | Apparatus and method for managing schedule of user of mobile terminal |
US20170024791A1 (en) | 2007-11-20 | 2017-01-26 | Theresa Klinger | System and method for interactive metadata and intelligent propagation for electronic multimedia |
US8516521B2 (en) * | 2008-01-14 | 2013-08-20 | Verizon Patent And Licensing Inc. | Interactive learning |
AU2008354378A1 (en) * | 2008-04-11 | 2009-10-15 | Gvbb Holdings S.A.R.L. | Method for automated television production |
US9246613B2 (en) * | 2008-05-20 | 2016-01-26 | Verizon Patent And Licensing Inc. | Method and apparatus for providing online social networking for television viewing |
US8548503B2 (en) * | 2008-08-28 | 2013-10-01 | Aol Inc. | Methods and system for providing location-based communication services |
US7974983B2 (en) | 2008-11-13 | 2011-07-05 | Buzzient, Inc. | Website network and advertisement analysis using analytic measurement of online social media content |
US8489515B2 (en) * | 2009-05-08 | 2013-07-16 | Comcast Interactive Media, LLC. | Social network based recommendation method and system |
KR20100123059A (en) | 2009-05-14 | 2010-11-24 | 삼성전자주식회사 | Contents information transmitting method, contents proposing method, reliability providing method of contents proposing information, and multimedia device using the same |
US8527602B1 (en) * | 2009-05-28 | 2013-09-03 | Google Inc. | Content upload system with preview and user demand based upload prioritization |
US9503496B2 (en) * | 2009-07-08 | 2016-11-22 | Centurylink Intellectual Property Llc | System and method for a media content reconciler |
KR102112973B1 (en) | 2009-07-16 | 2020-05-19 | 블루핀 랩스, 인코포레이티드 | Estimating and displaying social interest in time-based media |
US10339541B2 (en) * | 2009-08-19 | 2019-07-02 | Oracle International Corporation | Systems and methods for creating and inserting application media content into social media system displays |
US20110112899A1 (en) * | 2009-08-19 | 2011-05-12 | Vitrue, Inc. | Systems and methods for managing marketing programs on multiple social media systems |
US8510383B2 (en) * | 2009-09-14 | 2013-08-13 | Clixtr, Inc. | Method for providing event based media streams |
BR112012007674A2 (en) * | 2009-10-05 | 2016-11-29 | Fabtale Productions Pty Ltd | method and apparatus for providing an electronic document to a user, software for use with a computer, computer readable media, method for sharing content with one or more users for insertion into an electronic document, method and server for providing content to a or more users for insertion into an electronic document |
US9119027B2 (en) * | 2009-10-06 | 2015-08-25 | Facebook, Inc. | Sharing of location-based content item in social networking service |
US11122009B2 (en) | 2009-12-01 | 2021-09-14 | Apple Inc. | Systems and methods for identifying geographic locations of social media content collected over social networks |
US20130304818A1 (en) | 2009-12-01 | 2013-11-14 | Topsy Labs, Inc. | Systems and methods for discovery of related terms for social media content collection over social networks |
US9275119B2 (en) | 2010-04-19 | 2016-03-01 | Facebook, Inc. | Sharing search queries on online social network |
WO2011156633A2 (en) * | 2010-06-09 | 2011-12-15 | Brian Holmes | Live event social networking system |
JP2011259383A (en) * | 2010-06-11 | 2011-12-22 | Nippon Telegr & Teleph Corp <Ntt> | Tv-program-related information display system, interterminal event synchronizing apparatus, interterminal event synchronizing method, and program |
US20110320380A1 (en) * | 2010-06-23 | 2011-12-29 | Microsoft Corporation | Video content recommendations |
US20120036524A1 (en) * | 2010-08-03 | 2012-02-09 | Satish Mugulavalli | System and method for social networking using an internet television |
US10003857B2 (en) * | 2010-08-09 | 2018-06-19 | Surewaves Mediatech Private Limited | Method and system for inserting a local television content and a regional advertisement under centralized control |
US8627358B1 (en) * | 2010-08-16 | 2014-01-07 | West Corporation | Location-based movie identification systems and methods |
US20120047529A1 (en) * | 2010-08-20 | 2012-02-23 | Jennifer Schultz | Television and social network interworking system and method |
US20120110429A1 (en) * | 2010-09-23 | 2012-05-03 | Webdoc Sa | Platform enabling web-based interpersonal communication within shared digital media |
US20120084811A1 (en) * | 2010-10-04 | 2012-04-05 | Mark Thompson | System and Method for Integrating E-Commerce Into Real Time Video Content Advertising |
US20120084807A1 (en) * | 2010-10-04 | 2012-04-05 | Mark Thompson | System and Method for Integrating Interactive Advertising Into Real Time Video Content |
US8955001B2 (en) | 2011-07-06 | 2015-02-10 | Symphony Advanced Media | Mobile remote media control platform apparatuses and methods |
AU2011326430B2 (en) * | 2010-11-11 | 2015-01-22 | Google Llc | Learning tags for video annotation using latent subtags |
US8671433B2 (en) * | 2010-11-19 | 2014-03-11 | Media Patents, S.L. | Methods, apparatus and systems for delivering and receiving data |
KR101287110B1 (en) * | 2010-12-15 | 2013-07-17 | 주식회사 이니셜티 | Broadcasting receiving terminal, and control Method thereof |
US8863185B2 (en) * | 2011-01-04 | 2014-10-14 | Verizon Patent And Licensing Inc. | Social television |
US20120185892A1 (en) * | 2011-01-19 | 2012-07-19 | Fliptop, Inc., a corporation of CA | System and method for managing multiple content channels and engagement scoring |
US8898698B2 (en) * | 2011-01-21 | 2014-11-25 | Bluefin Labs, Inc. | Cross media targeted message synchronization |
US8943054B2 (en) | 2011-01-31 | 2015-01-27 | Social Resolve, Llc | Social media content management system and method |
US20120210351A1 (en) * | 2011-02-11 | 2012-08-16 | Microsoft Corporation | Presentation of customized digital media programming |
WO2012135804A2 (en) * | 2011-04-01 | 2012-10-04 | Mixaroo, Inc. | System and method for real-time processing, storage, indexing, and delivery of segmented video |
US9026596B2 (en) * | 2011-06-16 | 2015-05-05 | Microsoft Technology Licensing, Llc | Sharing of event media streams |
US10225617B2 (en) | 2011-06-22 | 2019-03-05 | Time Warner Cable Enterprises Llc | Methods and apparatus for presenting social network content in conjunction with video content |
US9681160B2 (en) * | 2011-06-22 | 2017-06-13 | Tout Inc. | Method and apparatus for automatically associating media segments with broadcast media streams |
US8473550B2 (en) * | 2011-09-21 | 2013-06-25 | Color Labs, Inc. | Content sharing using notification within a social networking environment |
EP2761882A4 (en) * | 2011-09-30 | 2015-05-27 | Tata Consultancy Services Ltd | A method and system for television program recommendation |
US20130132861A1 (en) * | 2011-11-22 | 2013-05-23 | Salesforce.Com, Inc. | Social media dashboards |
US9473809B2 (en) * | 2011-11-29 | 2016-10-18 | At&T Intellectual Property I, L.P. | Method and apparatus for providing personalized content |
US8930992B2 (en) * | 2012-01-19 | 2015-01-06 | Sony Corporation | TV social network advertising |
US10574711B2 (en) | 2012-02-09 | 2020-02-25 | Surewaves Mediatech Private Limited | Efficient multimedia content discovery and navigation based on reason for recommendation |
US9621932B2 (en) * | 2012-02-28 | 2017-04-11 | Google Inc. | Enhancing live broadcast viewing through display of filtered internet information streams |
US20130239146A1 (en) * | 2012-03-06 | 2013-09-12 | Aereo, Inc. | User Interface for Television Programming and Social Media Feeds for Time Shifted Viewing |
US10327032B2 (en) * | 2012-03-29 | 2019-06-18 | Sony Interactive Entertainment LLC | Extracting media content from social networking services |
US9301016B2 (en) * | 2012-04-05 | 2016-03-29 | Facebook, Inc. | Sharing television and video programming through social networking |
US20130268962A1 (en) | 2012-04-10 | 2013-10-10 | Shawn Andrew SNIDER | Integration of social media with live events |
CA2872051A1 (en) * | 2012-05-01 | 2013-11-07 | Agora Mobile Inc. | System and method for providing an application development and distribution social platform |
US20130297706A1 (en) * | 2012-05-03 | 2013-11-07 | United Video Properties, Inc. | Systems and methods for processing input from a plurality of users to identify a type of media asset segment |
US20130315562A1 (en) * | 2012-05-22 | 2013-11-28 | International Business Machines Corporation | Integrating Social Media with Recordation of Broadcast Data |
US20130332521A1 (en) * | 2012-06-07 | 2013-12-12 | United Video Properties, Inc. | Systems and methods for compiling media information based on privacy and reliability metrics |
US9584834B1 (en) * | 2012-06-25 | 2017-02-28 | Google Inc. | Video broadcasting with geolocation |
US20140012895A1 (en) * | 2012-07-01 | 2014-01-09 | 2C Corp. | Universal Social Networking Interface for One-Click Posting From Mobile Devices |
CN104145434B (en) * | 2012-08-17 | 2017-12-12 | 青岛海信国际营销股份有限公司 | The channel switch device of intelligent television |
US9699485B2 (en) * | 2012-08-31 | 2017-07-04 | Facebook, Inc. | Sharing television and video programming through social networking |
US20140068677A1 (en) * | 2012-08-30 | 2014-03-06 | Stevie Tv Ltd. | System and method for generating content channels |
US20140063259A1 (en) * | 2012-08-31 | 2014-03-06 | Ihigh.Com, Inc. | Method and system for video production |
WO2014041547A1 (en) * | 2012-09-13 | 2014-03-20 | Yevvo Entertainment Inc. | Live video broadcasting from a mobile device |
US9934368B2 (en) * | 2012-10-02 | 2018-04-03 | Banjo, Inc. | User-generated content permissions status analysis system and method |
US9817997B2 (en) * | 2014-12-18 | 2017-11-14 | Banjo, Inc. | User-generated content permissions status analysis system and method |
US20160034712A1 (en) * | 2012-10-02 | 2016-02-04 | Banjo, Inc. | System and method for event-related content discovery, curation, and presentation |
US20150242518A1 (en) * | 2012-10-04 | 2015-08-27 | Debbie Rosenbaum | Systems and methods for closed loop confirmation of user generated content |
US20140108602A1 (en) * | 2012-10-13 | 2014-04-17 | Thomas Walter Barnes | Method and system for delivering time-sensitive, event-relevant interactive digital content to a user during a separate event being experienced by the user |
DE102013017085A1 (en) * | 2012-10-15 | 2014-04-17 | Wixpress Ltd. | System for deep linking and search engine support for websites integrating a third-party application and components |
US9306989B1 (en) | 2012-10-16 | 2016-04-05 | Google Inc. | Linking social media and broadcast media |
US10403042B2 (en) | 2012-11-06 | 2019-09-03 | Oath Inc. | Systems and methods for generating and presenting augmented video content |
US20140136554A1 (en) * | 2012-11-14 | 2014-05-15 | National Public Radio, Inc. | System and method for recommending timely digital content |
US9055021B2 (en) * | 2012-11-30 | 2015-06-09 | The Nielsen Company (Us), Llc | Methods and apparatus to monitor impressions of social media messages |
US9569467B1 (en) * | 2012-12-05 | 2017-02-14 | Level 2 News Innovation LLC | Intelligent news management platform and social network |
US20140173644A1 (en) * | 2012-12-14 | 2014-06-19 | The Talent Channels Jv Operating Co Limited | Interactive celebrity portal and methods |
US20150088988A1 (en) * | 2012-12-21 | 2015-03-26 | Google Inc. | Social Queue on Television |
US9235625B2 (en) * | 2013-02-25 | 2016-01-12 | Google Inc. | Ranking media content sources |
US8850531B1 (en) * | 2013-03-07 | 2014-09-30 | Geofeedia, Inc. | System and method for targeted messaging, workflow management, and digital rights management for geofeeds |
US9049386B1 (en) * | 2013-03-14 | 2015-06-02 | Tribune Broadcasting Company, Llc | Systems and methods for causing a stunt switcher to run a bug-overlay DVE |
US9191422B2 (en) * | 2013-03-15 | 2015-11-17 | Arris Technology, Inc. | Processing of social media for selected time-shifted multimedia content |
US20140280571A1 (en) * | 2013-03-15 | 2014-09-18 | General Instrument Corporation | Processing of user-specific social media for time-shifted multimedia content |
US10607299B2 (en) | 2013-03-15 | 2020-03-31 | Tomorrowish Llc | Displaying social media content |
US9646650B2 (en) * | 2013-05-28 | 2017-05-09 | Google Inc. | Automatically syncing recordings between two or more content recording devices |
US9467750B2 (en) * | 2013-05-31 | 2016-10-11 | Adobe Systems Incorporated | Placing unobtrusive overlays in video content |
US9986307B2 (en) * | 2013-07-19 | 2018-05-29 | Bottle Rocket LLC | Interactive video viewing |
US9430532B2 (en) * | 2013-07-30 | 2016-08-30 | NETFLIX Inc. | Media content rankings for discovery of novel content |
US20150046842A1 (en) | 2013-08-08 | 2015-02-12 | Diageo North America, Inc. | System for providing a social media compilation |
KR20150018125A (en) * | 2013-08-09 | 2015-02-23 | 엘지전자 주식회사 | Electronic device and terminal communicating whit it |
US20150058448A1 (en) * | 2013-08-21 | 2015-02-26 | Josh Proctor | Internet video streaming system |
US8826347B1 (en) * | 2013-08-22 | 2014-09-02 | aioTV, Inc. | Method and system for creating video channels for delivery of content from multiple platforms |
US20150074728A1 (en) * | 2013-09-10 | 2015-03-12 | Opentv, Inc. | Systems and methods of displaying content |
US9953079B2 (en) * | 2013-09-17 | 2018-04-24 | International Business Machines Corporation | Preference based system and method for multiple feed aggregation and presentation |
US9798828B2 (en) * | 2013-10-25 | 2017-10-24 | Turner Broadcasting System, Inc. | Concepts for providing an enhanced media presentation |
US9319727B2 (en) | 2013-10-29 | 2016-04-19 | Fx Networks, Llc | Viewer-authored content acquisition and management system for in-the-moment broadcast in conjunction with media programs |
US9485543B2 (en) * | 2013-11-12 | 2016-11-01 | Google Inc. | Methods, systems, and media for presenting suggestions of media content |
US9552395B2 (en) * | 2013-11-13 | 2017-01-24 | Google Inc. | Methods, systems, and media for presenting recommended media content items |
US10104022B2 (en) * | 2013-11-15 | 2018-10-16 | Google Llc | Messaging for event live-stream |
US10349140B2 (en) * | 2013-11-18 | 2019-07-09 | Tagboard, Inc. | Systems and methods for creating and navigating broadcast-ready social content items in a live produced video |
US9742907B2 (en) * | 2013-12-05 | 2017-08-22 | Esna Technologies Inc. | System and method of providing social caller ID and caller rating on web calls |
US20150271557A1 (en) * | 2014-03-24 | 2015-09-24 | Joseph Akwo Tabe | Multimedia television system for interactive social media and social network |
US9992246B2 (en) * | 2014-03-27 | 2018-06-05 | Tvu Networks Corporation | Methods, apparatus, and systems for instantly sharing video content on social media |
US10327034B2 (en) * | 2014-03-27 | 2019-06-18 | Tvu Networks Corporation | Methods, apparatus and systems for exchange of video content |
US20150288645A1 (en) * | 2014-04-03 | 2015-10-08 | Avid Technology, Inc. | Synchronized story-centric media distribution |
US9519644B2 (en) * | 2014-04-04 | 2016-12-13 | Facebook, Inc. | Methods and devices for generating media items |
US9414011B2 (en) * | 2014-05-07 | 2016-08-09 | Disney Enterprises, Inc. | Remote, directed delivery of data for on-air graphics |
US20160057457A1 (en) * | 2014-08-25 | 2016-02-25 | LiveFromMe | Live video broadcasting mobile application for social sharing |
US20160055164A1 (en) * | 2014-08-25 | 2016-02-25 | Tll, Llc | News alert system and method |
US10652127B2 (en) | 2014-10-03 | 2020-05-12 | The Nielsen Company (Us), Llc | Fusing online media monitoring data with secondary online data feeds to generate ratings data for online media exposure |
WO2016081856A1 (en) * | 2014-11-21 | 2016-05-26 | Whip Networks, Inc. | Media management and sharing system |
US9521515B2 (en) * | 2015-01-26 | 2016-12-13 | Mobli Technologies 2010 Ltd. | Content request by location |
EP3629587A1 (en) * | 2015-03-27 | 2020-04-01 | Twitter, Inc. | Live video streaming services |
US10431208B2 (en) * | 2015-06-01 | 2019-10-01 | Sinclair Broadcast Group, Inc. | Content presentation analytics and optimization |
US10798193B2 (en) * | 2015-06-03 | 2020-10-06 | Oath Inc. | System and method for automatic storyline construction based on determined breaking news |
US9883249B2 (en) * | 2015-06-26 | 2018-01-30 | Amazon Technologies, Inc. | Broadcaster tools for interactive shopping interfaces |
WO2017015098A1 (en) * | 2015-07-17 | 2017-01-26 | Tribune Broadcasting Company, Llc | Video-production system with social-media features |
US9866923B2 (en) * | 2015-08-28 | 2018-01-09 | Cbs Interactive Inc. | Digital channel integration system |
US9635079B1 (en) * | 2015-11-12 | 2017-04-25 | International Business Machines Corporation | Social media sharing based on video content |
US10284883B2 (en) * | 2016-09-30 | 2019-05-07 | Disney Enterprises, Inc. | Real-time data updates from a run down system for a video broadcast |
US10405018B2 (en) * | 2017-11-06 | 2019-09-03 | Qvc, Inc. | System and method for publishing content for broadcast |
-
2016
- 2016-07-15 WO PCT/US2016/042456 patent/WO2017015098A1/en active Application Filing
- 2016-07-15 US US15/211,153 patent/US10291679B2/en not_active Expired - Fee Related
- 2016-07-15 WO PCT/US2016/042411 patent/WO2017015082A1/en active Application Filing
- 2016-07-15 WO PCT/US2016/042432 patent/WO2017015090A1/en active Application Filing
- 2016-07-15 WO PCT/US2016/042503 patent/WO2017015112A1/en active Application Filing
- 2016-07-15 WO PCT/US2016/042408 patent/WO2017015081A1/en active Application Filing
- 2016-07-15 EP EP16828296.0A patent/EP3326377A4/en not_active Withdrawn
- 2016-07-15 WO PCT/US2016/042461 patent/WO2017015100A1/en active Application Filing
- 2016-07-15 CA CA2992484A patent/CA2992484A1/en not_active Abandoned
- 2016-07-15 US US15/211,252 patent/US10412142B2/en not_active Expired - Fee Related
- 2016-07-15 US US15/211,170 patent/US20170019450A1/en not_active Abandoned
- 2016-07-15 US US15/211,243 patent/US20170017371A1/en not_active Abandoned
- 2016-07-15 EP EP16828297.8A patent/EP3326378A4/en not_active Withdrawn
- 2016-07-15 US US15/211,165 patent/US20170019694A1/en not_active Abandoned
- 2016-07-15 US US15/211,149 patent/US9979770B2/en not_active Expired - Fee Related
- 2016-07-15 CA CA2992495A patent/CA2992495A1/en not_active Abandoned
- 2016-07-15 US US15/211,155 patent/US20170019448A1/en not_active Abandoned
- 2016-07-15 US US15/211,189 patent/US10320876B2/en not_active Expired - Fee Related
- 2016-07-15 CA CA2992471A patent/CA2992471A1/en not_active Abandoned
- 2016-07-15 WO PCT/US2016/042505 patent/WO2017015114A1/en active Application Filing
- 2016-07-15 US US15/211,193 patent/US9961123B2/en not_active Expired - Fee Related
- 2016-07-18 EP EP16828375.2A patent/EP3326380A4/en not_active Withdrawn
- 2016-07-18 US US15/212,634 patent/US10305957B2/en not_active Expired - Fee Related
- 2016-07-18 CA CA2991623A patent/CA2991623A1/en not_active Abandoned
- 2016-07-18 CA CA2991631A patent/CA2991631A1/en not_active Abandoned
- 2016-07-18 US US15/213,126 patent/US20170019715A1/en not_active Abandoned
- 2016-07-18 WO PCT/US2016/042828 patent/WO2017015223A1/en active Application Filing
- 2016-07-18 US US15/213,123 patent/US20170019697A1/en not_active Abandoned
- 2016-07-18 EP EP16828355.4A patent/EP3326379A4/en not_active Withdrawn
- 2016-07-18 WO PCT/US2016/042764 patent/WO2017015197A1/en active Application Filing
-
2018
- 2018-01-08 IL IL256782A patent/IL256782A/en unknown
- 2018-01-08 IL IL256781A patent/IL256781A/en unknown
- 2018-01-14 IL IL256887A patent/IL256887A/en unknown
- 2018-01-14 IL IL256886A patent/IL256886A/en unknown
- 2018-03-21 US US15/927,904 patent/US10348794B2/en not_active Expired - Fee Related
-
2019
- 2019-03-27 US US16/366,527 patent/US20190222624A1/en not_active Abandoned
- 2019-04-16 US US16/386,242 patent/US10581947B2/en not_active Expired - Fee Related
- 2019-04-29 US US16/397,739 patent/US20190253474A1/en not_active Abandoned
- 2019-06-25 US US16/452,022 patent/US20190320003A1/en not_active Abandoned
- 2019-08-05 US US16/531,978 patent/US20190364085A1/en not_active Abandoned
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060156219A1 (en) * | 2001-06-27 | 2006-07-13 | Mci, Llc. | Method and system for providing distributed editing and storage of digital media over a network |
US20030122863A1 (en) * | 2001-12-28 | 2003-07-03 | International Business Machines Corporation | Navigation tool for slide presentations |
US20030149975A1 (en) * | 2002-02-05 | 2003-08-07 | Charles Eldering | Targeted advertising in on demand programming |
US20070168543A1 (en) * | 2004-06-07 | 2007-07-19 | Jason Krikorian | Capturing and Sharing Media Content |
US20070106419A1 (en) * | 2005-09-07 | 2007-05-10 | Verizon Business Network Services Inc. | Method and system for video monitoring |
US7733808B2 (en) * | 2006-11-10 | 2010-06-08 | Microsoft Corporation | Peer-to-peer aided live video sharing system |
US8539542B1 (en) * | 2009-08-25 | 2013-09-17 | Whdc Llc | System and method for managing multiple live video broadcasts via a public data network on a single viewing channel |
US20120192225A1 (en) * | 2011-01-25 | 2012-07-26 | Youtoo Technologies, LLC | Administration of Content Creation and Distribution System |
US20130104177A1 (en) * | 2011-10-19 | 2013-04-25 | Google Inc. | Distributed real-time video processing |
US20140341527A1 (en) * | 2013-05-15 | 2014-11-20 | MixBit, Inc. | Creating, Editing, and Publishing a Video Using a Mobile Device |
US20150215665A1 (en) * | 2014-01-30 | 2015-07-30 | Echostar Technologies L.L.C. | Methods and apparatus to synchronize second screen content with audio/video programming using closed captioning data |
US20150326814A1 (en) * | 2014-05-12 | 2015-11-12 | Echostar Uk Holdings Limited | Systems and method for timing commercial breaks |
US20160073029A1 (en) * | 2014-09-07 | 2016-03-10 | Guy MARKOVITZ | Method and system for creating a video |
US20160182979A1 (en) * | 2014-12-22 | 2016-06-23 | Sony Corporation | Method and system for processing closed-caption information |
US20160300594A1 (en) * | 2015-04-10 | 2016-10-13 | OMiro IP LLC | Video creation, editing, and sharing for social media |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10021442B1 (en) * | 2013-03-14 | 2018-07-10 | Tribune Broadcasting Company, Llc | Systems and methods for causing a stunt switcher to run a bug-removal DVE |
US10104449B1 (en) | 2013-03-14 | 2018-10-16 | Tribune Broadcasting Company, Llc | Systems and methods for causing a stunt switcher to run a bug-overlay DVE |
US10341737B2 (en) * | 2016-09-23 | 2019-07-02 | DISH Technologies L.L.C. | Integrating broadcast media streams with user media streams |
US20190273970A1 (en) * | 2016-09-23 | 2019-09-05 | DISH Technologies L.L.C. | Integrating broadcast media streams with user media streams |
US10880612B2 (en) * | 2016-09-23 | 2020-12-29 | DISH Technologies L.L.C. | Integrating broadcast media streams with user media streams |
US20210120310A1 (en) * | 2016-09-23 | 2021-04-22 | DISH Technologies L.L.C. | Integrating broadcast media streams with user media streams |
US11659251B2 (en) * | 2016-09-23 | 2023-05-23 | DISH Technologies L.L.C. | Integrating broadcast media streams with user media streams |
US20220321969A1 (en) * | 2021-04-02 | 2022-10-06 | Shanghai Bilibili Technology Co., Ltd. | Bullet comment processing method and apparatus |
US11812110B2 (en) * | 2021-04-02 | 2023-11-07 | Shanghai Bilibili Technology Co., Ltd. | Bullet comment processing method and apparatus |
Also Published As
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10581947B2 (en) | Video production system with DVE feature | |
US10381043B2 (en) | Media-production system with social media content interface feature | |
US20190327531A1 (en) | Video Production System with Content Extraction Feature | |
US10037780B1 (en) | Computing system with video content generation feature |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: TRIBUNE BROADCASTING COMPANY, LLC, ILLINOIS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HUNDEMER, HANK J.;REEL/FRAME:039182/0046 Effective date: 20160718 |
|
AS | Assignment |
Owner name: JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT, ILLINOIS Free format text: SECURITY AGREEMENT;ASSIGNORS:GRACENOTE, INC.;CASTTV, INC.;TRIBUNE BROADCASTING COMPANY, LLC;REEL/FRAME:039667/0565 Effective date: 20160809 Owner name: JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT Free format text: SECURITY AGREEMENT;ASSIGNORS:GRACENOTE, INC.;CASTTV, INC.;TRIBUNE BROADCASTING COMPANY, LLC;REEL/FRAME:039667/0565 Effective date: 20160809 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
AS | Assignment |
Owner name: BANK OF AMERICA, N.A., NORTH CAROLINA Free format text: SECURITY AGREEMENT;ASSIGNORS:CHICAGOLAND TELEVISION NEWS, INC.;KPLR, INC.;KSTU, LLC;AND OTHERS;REEL/FRAME:050438/0004 Effective date: 20190919 |
|
AS | Assignment |
Owner name: TRIBUNE BROADCASTING COMPANY, LLC, ILLINOIS Free format text: PARTIAL RELEASE - REEL/FRAME 039667/0565;ASSIGNOR:JPMORGAN CHASE BANK, N.A., AS COLLATERAL AGENT;REEL/FRAME:050474/0975 Effective date: 20190919 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |