Multimedia content preview rendering in a cloud content management system

Information

  • Patent Grant
  • 9063912
  • Patent Number
    9,063,912
  • Date Filed
    Wednesday, June 22, 2011
    13 years ago
  • Date Issued
    Tuesday, June 23, 2015
    9 years ago
Abstract
Systems and methods for providing content previews of files, including video files, stored in an online content management system with collaboration environment are disclosed. Permission to view a preview of a file is determined based upon access permissions associated with the file and/or workspace in which the file is stored and the access level of the requesting user. The system can also provide a progress indicator indicating the approximate time remaining to process the file for previewing based upon averages of times for processing previous files having similar attributes. In one aspect, embodiments of the present disclosure include a method, which may be implemented on a system, for downsampling a video file to decrease the processing time for providing a preview.
Description
CROSS-REFERENCES TO RELATED APPLICATIONS

This application is related to co-pending U.S. application Ser. No. 13/152,982, entitled, “REAL TIME NOTIFICATION OF ACTIVITIES THAT OCCUR IN A WEB-BASED COLLABORATION ENVIRONMENT”, filed Jun. 3, 2011 and is incorporated in its entirety.


BACKGROUND

Online file storage systems typically provide a way for a user to upload files to a storage server for backing up files, file access, and/or file distribution. Some online file storage systems allow the user to preview the contents of a file before or instead of downloading the file.





BRIEF DESCRIPTION OF THE DRAWINGS

Examples of a web-based content management system with a collaboration environment that provides previewing of files are illustrated in the figures. The examples and figures are illustrative rather than limiting.



FIG. 1 illustrates an example diagram of a system where a host server and conversion servers support previewing of files in an online collaboration environment.



FIG. 2 depicts an example diagram of a web-based or online collaboration platform deployed in an enterprise or other organizational setting for organizing work items and workspaces.



FIG. 3A depicts a block diagram illustrating an example of components in the host server of a web-based collaboration environment that supports previewing of files.



FIG. 3B depicts a block diagram illustrating an example of components in the host server of a web-based collaboration environment with real time activity notification capabilities.



FIG. 3C depicts a block diagram illustrating an example of components in a notification server for providing real time or near real time notifications of activities that occur in a web-based or online collaboration environment.



FIG. 4 depicts a block diagram illustrating an example of components in a conversion server for providing conversion of a video file format into a file format suitable for previewing the file in a web-based or online collaboration environment.



FIG. 5 depicts a flow chart illustrating an example process for determining if a user has permission to access a file preview.



FIG. 6 depicts a flow chart illustrating an example process for finding an optimum path using different conversion tools for converting an input file format to a target file format.



FIG. 7 depicts a flow chart illustrating an example process for converting a format of a video file to a target format for previewing.



FIG. 8 depicts a flow chart illustrating an example process for generating a progress indicator for indicating the amount of time remaining until a file has been converted.



FIG. 9 shows a diagrammatic representation of a machine in the example form of a computer system within which a set of instructions, for causing the machine to perform any one or more of the methodologies discussed herein, may be executed.



FIG. 10 depicts a screenshot showing an example of a user interface of a file page with a feed stream of real time or near real time notifications of activities that occur in a web-based collaboration environment.



FIG. 11 depicts a screenshot showing an example of a user interface of a preview page with a feed stream of real time or near real time notifications of activities that occur in a web-based collaboration environment.





DETAILED DESCRIPTION

A system is described for providing content previews of files stored in an online content management system with a collaboration environment. Each collaborator of a workspace in the collaboration environment is assigned a permission level or has a default permission level. Permission to view a preview of a file is determined based upon access permissions associated with the file and/or workspace in which the file is stored and the access level of the requesting user.


Further, the system can provide a visual indicator, such as a progress bar, that indicates the time remaining to process a file for previewing. The anticipated duration for processing a file is based upon averages of times needed to process previous files having similar attributes. Additionally, embodiments of the present disclosure include a method for downsampling a video file to decrease the processing time for providing a preview.


Various aspects and examples of the invention will now be described. The following description provides specific details for a thorough understanding and enabling description of these examples. One skilled in the art will understand, however, that the invention may be practiced without many of these details. Additionally, some well-known structures or functions may not be shown or described in detail, so as to avoid unnecessarily obscuring the relevant description.


The terminology used in the description presented below is intended to be interpreted in its broadest reasonable manner, even though it is being used in conjunction with a detailed description of certain specific examples of the technology. Certain terms may even be emphasized below; however, any terminology intended to be interpreted in any restricted manner will be overtly and specifically defined as such in this Detailed Description section.



FIG. 1 illustrates an example diagram of a system where a host server 100 and conversion servers 150 support previewing of files for users 108 in an online collaboration environment so that the users 108 do not have to download the file to access its contents. Further, the host server 100 and notification server 150 provide notifications of activities that occur in the online collaboration environment in ire realtime or near real time to users 108.


The client devices 102 can be any system and/or device, and/or any combination of devices/systems that is able to establish a connection, including wired, wireless, cellular connections with another device, a server and/or other systems such as host server 100 and/or conversion servers 150. Client devices 102 will typically include a display and/or other output functionalities to present information and data exchanged between or among the devices 102 and/or the host server 100 and/or conversion servers 150. In one embodiment, this is only a single conversion server 150. In one embodiment, there are multiple conversion servers 150 working independently. In other embodiments, a distributed queuing system is used to run the conversion servers 150 in an efficient manner.


For example, the client devices 102 can include mobile, hand held or portable devices or non-portable devices and can be any of, but not limited to, a server desktop, a desktop computer, a computer cluster, or portable devices including, a notebook, a laptop computer, a handheld computer, a palmtop computer, a mobile phone, a cell phone, a smart phone, a PDA, a Blackberry device, a Treo, a handheld tablet (e.g. an iPad, a Galaxy, Xoom Tablet, etc.), a tablet PC, a thin-client, a hand held console, a hand held gaming device or console, an iPhone, and/or any other portable, mobile, hand held devices, etc. In one embodiment, the client devices 102, host server 100, and conversion servers 150 are coupled via a network 106 and/or a network 108. In some embodiments, the devices 102 and host server 100 may be directly connected to one another.


The input mechanism on client devices 102 can include touch screen keypad (including single touch, multi-touch, gesture sensing in 2D or 3D, etc.), a physical keypad, a mouse, a pointer, a track pad, motion detector (e.g., including 1-axis, 2-axis, 3-axis accelerometer, etc.), a light sensor, capacitance sensor, resistance sensor, temperature sensor, proximity sensor, a piezoelectric device, device orientation detector (e.g., electronic compass, tilt sensor, rotation sensor, gyroscope, accelerometer), or a combination of the above.


Signals received or detected indicating user activity at client devices 102 through one or more of the above input mechanism, or others, can be used in the disclosed technology by various users or collaborators (e.g., collaborators 108) for accessing, through network 106, a web-based collaboration environment or online collaboration platform (e.g., hosted by the host server 100).


The collaboration platform or environment hosts workspaces with work items that one or more users can access (e.g., view, edit, update, revise, comment, discussions, download, preview, tag, or otherwise manipulate, etc.). A work item can generally include any type of digital or electronic content that can be viewed or accessed via an electronic device (e.g., device 102). The digital content can include .pdf files, .doc, slides (e.g., Powerpoint slides), images, audio files, multimedia content, web pages, blogs, etc. A workspace can generally refer to any grouping of a set of digital content in the collaboration platform. The grouping can be created, identified, or specified by a user or through other means. This user may be a creator user or administrative user, for example.


In general, a workspace can be associated with a set of users or collaborators (e.g., collaborators 108) which have access to the content included therein. The levels of access (e.g., based on permissions or rules) of each user or collaborator to access the content in a given workspace may be the same or may vary among the users. Each user may have their own set of access rights to every piece of content in the workspace, or each user may have different access rights to different pieces of content. Access rights may be specified by a user associated with a work space and/or a user who created/uploaded a particular piece of content to the workspace, or any other designated user or collaborator.


In general, the collaboration platform allows multiple users or collaborators to access or collaborate efforts on work items such that each user can see, remotely, edits, revisions, comments, or annotations being made to specific work items through their own user devices. For example., a user can upload a document to a work space for other users to access (e.g., for viewing, editing, commenting, discussing, signing-off, or otherwise manipulating). The user can login to the online platform and upload the document (or any other type of work item) to an existing work space or to a new work space. The document can be shared with existing users or collaborators in a work space.


A diagrammatic illustration of the online collaboration environment and the relationships between workspaces and users/collaborators are illustrated with further reference to the example of FIG. 2.


In one embodiment, actions performed on work items or other activities that occur in a work space can be detected in real time or in near real time. In addition, users, collaborators, or select users can be notified in real time or near real-time of these actions or activities. Various mechanisms can be used to notify users or collaborators, including through the web interface to access the collaboration platform, via email, and/or SMS, for example.


Functions and techniques disclosed for previewing of files or work items in the online platform can be performed by one or more distributed conversion servers 150 of the collaboration platform. Functions and techniques performed by the host server 100, the conversion servers 150, and the related components therein are described, respectively, in detail with further reference to the examples of FIGS. 3A and 4.


In one embodiment, client devices 102 communicate with the host server 100 and/or conversion servers 150 over network 106. In general, network 106, over which the client devices 102, the host server 100, and/or conversion servers 150 communicate, may be a cellular network, a telephonic network, an open network, such as the Internet, or a private network, such as an intranet and/or the extranet, or any combination thereof. For example, the Internet can provide file transfer, remote log in, email, news, RSS, cloud-based services, instant messaging, visual voicemail, push mail, VoIP, and other services through any known or convenient protocol, such as, but is not limited to the TCP/IP protocol, Open System Interconnections (OSI), FTP, UPnP, iSCSI, NSF, ISDN, PDH, RS-232, SDH, SONET, etc.


The network 106 can be any collection of distinct networks operating wholly or partially in conjunction to provide connectivity to the client devices 102 and the host server 100 and may appear as one or more networks to the serviced systems and devices. In one embodiment, communications to and from the client devices 102 can be achieved by, an open network, such as the Internet, or a private network, such as an intranet and/or the extranet. In one embodiment, communications can be achieved by a secure communications protocol, such as secure sockets layer (SSL), or transport layer security (TLS).


In addition, communications can be achieved via one or more networks, such as, but are not limited to, one or more of WiMax, a Local Area Network (LAN), Wireless Local Area Network (WLAN), a Personal area network (PAN), a Campus area network (CAN), a Metropolitan area network (MAN), a Wide area network (WAN), a Wireless wide area network (WWAN), enabled with technologies such as, by way of example, Global System for Mobile Communications (GSM), Personal Communications Service (PCS), Digital Advanced Mobile Phone Service (D-Amps), Bluetooth, Wi-Fi, Fixed Wireless Data, 2G, 2.5G, 3G, 4G, IMT-Advanced, pre-4G, 3G LTE, 3GPP LTE, LTE Advanced, mobile WiMax, WiMax 2, WirelessMAN-Advanced networks, enhanced data rates for GSM evolution (EDGE), General packet radio service (GPRS), enhanced GPRS, iBurst, UMTS, HSPDA, HSUPA, HSPA, UMTS-TDD, 1xRTT, EV-DO, messaging protocols such as, TCP/IP, SMS, MMS, extensible messaging and presence protocol (XMPP), real time messaging protocol (RTMP), instant messaging and presence protocol (IMPP), instant messaging, USSD, IRC, or any other wireless data networks or messaging protocols.



FIG. 2 depicts an example diagram of a web-based or online collaboration platform deployed in an enterprise or other organizational setting 250 for organizing work items 215, 235, 255 and workspaces 205, 225, 245.


The web-based platform for collaborating on projects or jointly working on documents can be used by individual users and shared among collaborators. In addition, the collaboration platform can be deployed in an organized setting including but not limited to, a company (e.g., an enterprise setting), a department in a company, an academic institution, a department in an academic institution, a class or course setting, or any other types of organizations or organized setting.


When deployed in an organizational setting, multiple workspaces (e.g., workspace A, B C) can be created to support different projects or a variety of work flows. Each workspace can have its own associated work items. For example, work space A 205 may be associated with work items 215, work space B 225 can be associated with work items 235, and work space N 245 can be associated with work items 255. The work items 215, 235, and 255 may be unique to each work space but need not be. For example, a particular word document can be associated with only one work space (e.g., work space A 205) or it may be associated with multiple work spaces (e.g., Work space A 205 and work space B 225, etc.).


In general, each work space has a set of users or collaborators associated with it. For example, work space A 205 is associated with multiple users or collaborators 206. In some instances, work spaces deployed in an enterprise may be department specific. For example, work space B may be associated with department 210 and some users shown as example user A 208 and workspace N 245 can be associated with departments 212 and 216 and users shown as example user B 214.


Each user associated with a work space can generally access the work items associated with the work space. The level of access will depend on permissions associated with the specific work space, and/or with a specific work item. Permissions can be set for the work space or set individually on a per work item basis. For example, the creator of a work space (e.g., one of user A 208 who creates work space B) can set one permission setting applicable to all work items 235 for other associated users and/or users associated with the affiliate department 210, for example. Creator user A 208 may also set different permission settings for each work item, which may be the same for different users, or varying for different users.


In each work space A, B . . . N, when an action is performed on a work item by a given user or any other activity is detected in the work space, other users in the same work space may be notified in real time or in near real time. Activities which trigger real time notifications can include, by way of example but not limitation, adding, deleting, or modifying collaborators in the work space, adding, deleting a work item in the work space, creating a discussion topic in the work space.


The activity can be performed in relation to a discussion topic in the work space, for example, adding a response to a discussion topic, deleting a response, or editing a response in the work space. In addition, the activity is performed on a work item in the work space by the user, including, by way of example but not limitation, download or upload of a work item, deletion of editing of the work item, selecting, adding, deleting, and modifying a tag in the work item, preview of the work item or comment of the work item, setting or changing permissions of the work item, sharing a work item, emailing a link to the work item, and/or embedding a link to the work item on another website.



FIG. 3A depicts a block diagram illustrating an example of components in the host server 100 of a web-based collaboration environment that supports previewing of files.


The host server 100 of the web-based or online collaboration environment can generally be a cloud-based service. The host server 100 can include, for example, a network interface 302, a load balancing engine 310, a network graph engine 320, an analytics database 336, and/or a progress indicator engine 330, and a data collection module 340. Additional or fewer components/modules/engines can be included in the host server 100 and each illustrated component.


The network interface 302 can be a networking module that enables the host server 100 to mediate data in a network with an entity that is external to the host server 100, through any known and/or convenient communications protocol supported by the host and the external entity. The network interface 302 can include one or more of a network adaptor card, a wireless network interface card (e.g., SMS interface, WiFi interface, interfaces for various generations of mobile communication standards including but not limited to 1G, 2G, 3G, 3.5G, 4G, LTE, etc.,), Bluetooth, a router, an access point, a wireless router, a switch, a multilayer switch, a protocol converter, a gateway, a bridge, bridge router, a hub, a digital media receiver, and/or a repeater.


As used herein, a “module,” “a manager,” a “handler,” a “detector,” an “interface,” or an “engine” includes a general purpose, dedicated or shared processor and, typically, firmware or software modules that are executed by the processor. Depending upon implementation-specific or other considerations, the module, manager, handler, or engine can be centralized or its functionality distributed. The module, manager, handler, or engine can include general or special purpose hardware, firmware, or software embodied in a computer-readable. (storage) medium for execution by the processor. As used herein, a computer-readable medium or computer-readable storage medium is intended to include all mediums that are statutory (e.g., in the United States, under 35 U.S.C. 101), and to specifically exclude all mediums that are non-statutory in nature to the extent that the exclusion is necessary for a claim that includes the computer-readable (storage) medium to be valid. Known statutory computer-readable mediums include hardware (e.g., registers, random access memory (RAM), non-volatile (NV) storage, to name a few), but may or may not be limited to hardware.


One embodiment of the host server 100 includes the data collection module 340 which performs real time collection of statistics on every job that runs through the conversion servers 150 during the conversion process. Some file formats do not need to be converted to any intermediate formats before being converted to the target format. These file formats can be converted in a single step. Other file formats use one or more intermediate steps to convert to intermediate formats before finally being converted to the target format. The data collection module 340 records the timing information for every step of the file conversion process and stores the information in the analytics database 336. By collecting timing data for every job and analyzing the data to optimize the number of jobs processed per server in a given amount of time, the overall throughput of jobs can be maximized by the system.


One embodiment of the host server 100 includes the load balancing engine 310 which can evaluate the load on each of the distributed conversion servers 150 and use a distributed queuing system to determine which of the distributed conversion servers 150 to send the next job. In one embodiment, jobs are sent to a particular conversion server based upon a round-robin process where each job is sent to a different server until every server has received a job, and further jobs are sent to the servers in the same order. In one embodiment, video conversion jobs can be queued in a cluster separately from other conversion jobs because video conversion can take longer than conversion of other types of files. Then for each cluster of jobs, the round-robin process would be used among the servers. Additionally, other classes of files could also be queued in their own special clusters, as the need arises.


One embodiment of the host server 100 includes the network graph engine 320 which can identify and evaluate the available tools for converting files from an existing format to a target format and generate a network graph to determine the optimum tool or tools to use for the conversion process of a given file. The network graph engine 320 can include a network graph generator 322 and a network graph solver 324.


The network graph generator 322 maintains information on the tools available in the system for converting files. Each tool has a list of supported input formats and one or more output formats to which the tool can convert a file. Based upon the input and output formats, the network graph generator 322 generates a network graph. The network graph is made up of nodes that represent a tool that can be used to convert a particular input file format to a particular output file format. The nodes are then connected so that the output format of a first node is connected to a second node that takes the output format as an input format. Every conversion tool available to the system and every pair of supported input-output formats is assigned to a different node, and the appropriate connections between the nodes are made, thus generating a network graph of conversion tools.


The network graph generator 322 maintains a priority cost associated with each tool for each set of input to output conversion function. The priority cost can be based on factors including, but not limited to, how long a particular conversion tool takes to perform a conversion, the fidelity of the output file of the conversion tool, and system preferences for conversion of a file into a particular file format. In one embodiment, conversion tools that convert the same input file formats to the same output file formats but use a different algorithm, e.g. ffmpeg versus mencoder, can be maintained with different priority costs for different file formats. In one embodiment, specific versions of conversion tools can also be maintained because a particular version performs better for a desired input-output file format conversion. For example, with the swftools library, an older version provides higher quality results for particular conversion situations. As a result, upgrading to the latest version degrades the quality of the output file. Thus, a node can be associated with each installed version of the conversion tool with specified input, output, and priority costs.


Based upon the priority cost, the network graph generator 322 solves the network graph to find the optimum conversion scenario for a given input file format and target output file format. If during the conversion of a file a particular conversion tool fails to create a sufficient output file, the network graph generator 322 can remove the effected node(s), and the network graph solver 324 can re-solve the network graph to re-route the conversion of the file around the fault.


One embodiment of the host server 100 includes the progress indicator engine 330 which can generate a progress bar or other type of progress indicator for display to the user to let the user know the approximate amount of time remaining for a file conversion to complete and when a display of the desired file preview will be available. The progress indicator engine 330 can include a duration analysis module 332 and/or an indicator generator 334 and can access the analytics database 336.


The duration analysis module 332 accesses data collected and stored in the analytics database 336 about previous files that have been converted by the system. The analytics database 336 includes, but is not limited to, data such as the size of a file and the format of the file. For text files, the data includes the number of pages of the file. For image and video files, the data includes the width and height of the source image format and the target image format, for example in number of pixels. For video files, the data also includes the play length of the video. The database can also include information about the number of steps used to convert the input file format to the target file format, for example, the number of intermediate file formats that the input file format needs to be converted to before reaching the final target format. In one embodiment, the analytics database 336 can be made up of several separate databases. In one embodiment, the analytics database 336 can be external to the conversion server 150 and/or shared by the conversion servers 150.


The duration analysis module 332 can find previous files that have similar attributes as the file to be converted and calculate a predicted conversion time based on the conversion times of similar files that have been converted in the past. In one embodiment, the duration analysis module 332 takes an average of the conversion times of the similar files as a predicted conversion duration for the file to be converted. In one embodiment, the duration analysis module 332 can perform a more complex calculation on the collected analytics data, for example, a weighted average depending upon how similar a previous input file is to the present file to be converted, to predict the conversion time for the present file. In one embodiment, the predicted conversion duration can be further refined by the duration analysis module 332 based on any steps of the conversion process that operate linearly and can report its progress.


Further, the analytics database 336 can include data on the load on the conversion servers at the time the previous files were converted. In one embodiment, the conversion server can comprise a number of distributed servers. By determining the load on the conversion servers when a file is to be converted to a target format, the average of the conversion times of similar files can be adjusted to the current load of the servers.


The indicator generator module 334 generates the progress indicator for displaying an indication of the time that has elapsed since the start of the conversion of the file and the expected time remaining until the file has finished converting to the target format. The indicator generator module 334 includes a timer 335 for tracking the elapsed time from when the conversion process starts to when the conversion process has finished.


In some embodiments, a video file can start to be streamed to the user prior to completion of the conversion process. In this case, the timer 335 tracks the elapsed time from when the conversion process starts to the anticipated time when the video player can start reading data and showing video to the user before the remainder of the file has been converted to the target format.



FIG. 3B depicts a block diagram illustrating an example of components in the notification module 390 that is part of the host server 100 of a web-based collaboration environment with real time activity notification capabilities.


The notification module 390 can include, for example, an activity detector 405, a notification engine 415, a permissions manager 435 and/or a notification prioritizer 455. The notification engine 415 can include a notification server manager 416 and/or a recipient selection module, the permission manager 435 can include a workspace manager 426, a collaboration manager 438, an organization tracker 440 having an enterprise tracker 441, and/or a work item manager 442; the notification prioritizer 455 can further include a rules module 456, a chronological ordering engine 458, and/or a relevance assessor 460. Additional or less components/modules/engines can be included in the notification module 390 and each illustrated component.


One embodiment of the notification module 390 includes the activity detector 405 which can detect an activity in the web-based collaboration environment. The activity can be a triggering activity which causes select or relevant users to be notified of the occurrence, which in one embodiment, is in real time or near real-time.


The detected activity can be performed by a user or collaborator in a work space and can be performed on a work item or relating to a work item, for example, download or upload of the work item, previewing, commenting of a work item, deletion or editing of the work item, commenting on a work item, identifying, selecting, adding, deleting, saving, editing, and modifying a tag in the work item, setting or changing permissions of the work item, sharing the work item including, for example, emailing a link to the work item, embedding a link to the work item on another website.


The types of activities that can be detected can also relate to changes to a work space, such as adding, deleting, or modifying collaborators in the work space; changes to work items such as adding, deleting a work item in the work space; creating a discussion topic in the work space, adding a response to a discussion topic, deleting a response, or editing a response in the work space.


Detected activity in a work space that is performed by a user or otherwise occurring can trigger notifications to be sent out, for example, via the notification engine 415. The notification engine 415 can notify users, which can be collaborators of the user who performed the activity in the work space via one or more of many mechanisms, including but not limited to, email, SMS, voice-message, text-based message, RSS, feed, etc.


In one embodiment, the notification is depicted through a web-browser used by the other user to access the web-based collaboration environment, for access in real time or near real time to when the activity was performed by the user. When notifying a user in real time through a web-browser, the notification engine 415 can utilize a push-enabled service to ensure real time notification. In one embodiment, the notification is sent by a component or another server which implements push technology (e.g., the notification server 500 shown in the example of FIG. 3C). The push-enabled service can be implemented via long poll or HTTP streaming, for example, by the notification sever 500 or another component, device which may be internal to or external to the host server 100.


The notification module 390 can send a notification server an identification of the recipient to be notified and indicator of the activity to notify the recipient of. Use of an external push server, such as the notification server 550 is described with further reference to the example of FIG. 3C. The notification server 550 can be managed by the notification server manager 416 in the notification engine 415 which can communicate events to notify users in real-time via their browser interfaces. In one embodiment, the host server sends a notification server an identification of the recipient to be notified and indicator of the activity to notify the recipient of.


In general, recipients of an activity notification are selected based on criteria, for example, by the recipient selection module 418 of the notification engine 415. The criteria may be determined, for example, based on a work space in which the activity was performed in the online collaboration platform. Specifically, the criteria, is in one embodiment, determined based on permissions configured for the workspace, as managed, tracked, updated, implemented, revised, based by the permissions manager 435.


For example, the workspace can be associated with an enterprise and in such an instance, the criteria can specify that the recipient that is selected is an employee of the enterprise. Enterprise associations and affiliations can be managed by the organization tracker 440, for example; in some embodiments, enterprises and/or enterprise accounts can specifically be managed, tracked, monitored by the enterprise tracker 441. Permissions for the workspace can configured by a creator or administrative user of the workspace. The collaboration manager 438 can determine, track, and implement relationships, roles, and/or access levels of multiple users/collaborators. For example, users may be a general user, a creator of the work space, a creator of a work item, or an administrative user. The permissions for a work space can be configured by a user, creator, or the administrative user and is generally managed by the collaborations manager 438.


The criteria that are determined by work space permissions can be managed by the work space manager 436 in the permissions manager 435. The recipient selection module 418 can also determine the recipient selection criteria based on user affiliation with the workspace, including, one or more of, member, invited collaborator and collaborator in the workspace. Such user affiliation can be tracked and managed by, for example, the collaboration manger 438 of the permissions manager 435.


In one embodiment, the criteria are determined based on permissions associated with a work item on which the activity was performed in relation to in the workspace. Permissions associated with work items can be managed, tracked, updated, revised, or implemented, in one embodiment, by the work item manager 442. For example, the permissions associated with the work item can be set by, a creator of the work item or an administrative user of the work space. Each work space can include multiple work items where each of multiple work items has individually configurable permissions. The individually configured permissions can be determined by user roles and rights (e.g., as managed by the collaborations manager 438). The work item manager 442 can communicate with the collaboration manager 438 in setting, configuring, or re-configuring permissions associated with work items.


The notification of a triggering activity, can be presented to a selected user in the web-based or online collaboration environment such that the notification is accessible by the user in real time or near real time to when the triggering activity occurred. In one embodiment, the notification is presented via a user interface to the online collaboration platform, for example, when the recipient (e.g., selected recipient) is accessing the workspace (e.g., the same work space in which activity is detected) or when the recipient is accessing a different work space. Specifically, the real time or near real time notification can be presented to the user via the user interface if the user is online (e.g., online or otherwise logged into the web-based or online collaboration environment). Examples of real time notifications are illustrated with further reference to the examples of FIG. 10 and FIG. 11. The feed stream is shown in the lower right corner of the files page in FIG. 10 and the preview page in FIG. 11.


The notification engine 415 can determine the channel through which to notify selected users or recipients of activity. The channels that are used can include, indicators via a user interface to the online collaboration environment, SMS, audio message, text-based messages, email, desktop application, RSS, etc. The indicators presented via the user interface can include visual indicators (e.g., pop-up form including text and/or graphics), audio indicators, or any other types detectable by a user.


In one embodiment, the notification is presented in the user interface among other notifications in an order based a rule, which may be configurable by the recipient or another user. Such prioritization in presentation can be determined, managed, tracked, implemented, revised, or updated by the notification prioritizer 455, for example. The notification prioritizer 455 can present the notification in the user interface (e.g., as shown in the user interface of FIG. 10) among other notifications in an order based on a rule as determined by the rules module 456, for example.


The rule can indicate user preferences for notifications of activities based on one or more of, a type of activity that occurred and a user related to the activity. For example, a given user may explicitly or implicitly indicate preferences for activities or actions performed by specific other users or collaborators. A user may also indicate explicitly or implicitly preferences for types of activities that they wish to be notified of or not notified of. Users may also indicate that notifications for certain types of activities are to be prioritized other others. For example, a user may indicate that a notification for a ‘comment on’ activity is of a higher priority compared to a ‘edit’ activity.


In one embodiment, the notification is presented in the user interface among other notifications based on chronological order, for example as tracked or determined by the chronological ordering engine 458. For example, each notification can be depicted in the user interface based the time ordering when each associated triggering activity occurred. Notification of the most recently occurred activity can be depicted above or below other notifications, or in a location where most easily accessed by the recipient user.


In one embodiment, the notification is presented in the user interface among other notifications based on relevance to the recipient, for example, as determined, tracked, monitored, or implemented by the relevance assessor 460. The relevance to the recipient can be represented by, for example, contextual and temporal parameters. For example, contextual parameters provide metrics indicating the recipient's current activity in the online collaboration platform. Current activity can be any activity of the user that occurred within a certain time frame (e.g., within the last minute, within the last 5 minutes, within the last 10 minutes, for example). Activity of the user can include, a document that the user edited, viewed, downloaded, commented on, tagged, or otherwise accessed. Activity of the user can also include activities surrounding a workspace, including creation/modification of a workspace or attributes of a workspace, such as modification of collaborators, permissions, etc.


Temporal parameters can, for example, provide metrics indicating the recipient's activities in the online collaboration platform over a period of time, a frequency with which the recipient has accessed a work item with which the activity relates to, and/or a frequency with which the recipient has accessed the work space in which the activity was performed.



FIG. 3C depicts a block diagram illustrating an example of components in a notification server 550 for providing real time or near real time notifications of activities that occur in a web-based or online collaboration environment.


The notification server 550 generally includes, for example, a push server 505, an SMS notifier 515, and/or a priority module 525. In one embodiment, the push server 505 includes a long poll engine 506 and/or an HTTP streaming engine 508. Additional or less components/modules/engines can be included in the notification server 550 and each illustrated component.


The notification server 550 can support the services of a collaboration platform or environment to provide real time or near real time notifications of activities. In one embodiment, the notification server 550 is integrated within a host server of a collaboration platform (e.g., the host server 100 shown in the example of FIG. 1). The notification server 550 may also be externally coupled to the host server (e.g., the host server 100). In some instances, a portion of the functions implemented and performed by the notification server 550 can be implemented in part or in whole in the host server 100. For example, some of the components shown to be in the notification server 500 and associated functionalities can in part or in whole reside in the host server 100.


In one embodiment, the notification server 550 sends a notification of an activity that occurs within a collaboration platform to a recipient. The notification is sent by the server 550 such that the recipient is notified in real time or near real time to when the activity occurred or when the activity was performed. Real time notification can be performed via push technology, for example by the push server 505 through long polls (e.g., via the long poll engine 506) and/or through the HTTP streaming (e.g., via the HTTP streaming engine 506). The notification server 550 can communicate with the host server to determine a recipient to whom to notify. The notification server 550 can also determine the activity to notify the recipient of, for example through communication with the host server.


In one embodiment, the notification is presented in a feed stream among other notifications through a user interface on the user device according to relevancy to the user determined based on current or recent activity of the user in the web-based collaboration environment. The presentation priority in a user interface in a feed stream can be managed, in whole, or in part, for example, by the priority module 525 using information determined by the notification prioritizer (e.g., notification prioritizer 455 of notification module 390 shown in the example of FIG. 3B).


In one embodiment, the notification server 550 can send notifications to users via SMS (e.g., through the SMS notifier 515). In this instance, the notification server 500 can be coupled to an SMS center which forwards the SMS text message to a mobile device over a cellular network. The notification can be sent via SMS in real time or near real time, or with a delay.



FIG. 4 depicts a block diagram illustrating an example of components in a conversion server 400 for converting a video file format into a target file format suitable for previewing the file in a web-based or online collaboration environment. The conversion server 400 of the web-based or online collaboration environment can include, for example, a conversion engine 410, a streaming engine 430, a downsampling engine 444, a fast start preview module 450, and/or a rotation module 466.


One embodiment of the conversion server 400 includes the conversion engine 410 which can run the conversion tools used to convert an input file format to another file format. Non-limiting examples of conversion tools include mencoder and ffmpeg. In one embodiment, the conversion tool manager 410 manages subsystems that are designed to handle specific style formats. For example, because Microsoft Office is ideally suited to convert MSWord and PowerPoint documents, a Windows-based subsystem can be used to handle MS Office file formats.


In one embodiment, the conversion engine 410 can perform a hard-coded conversion process where specific tools are used to convert certain input formats to a specified output format. In one embodiment, the conversion engine 410 uses a path-finding algorithm implemented by the network graph engine 320 described above.


One embodiment of the conversion server 400 includes the streaming engine 430 which can take the video file as it is being converted by the conversion engine 410 and write it directly to a publicly-available location in chunks. The video player can then start reading the data showing the video to the user before the rest of the file has been converted.


One embodiment of the conversion server 400 includes the downsampling engine 444 which can downsample a video file by removing pixels from video frames to reduce the size of the file for converting to a target format for preview. Advantages of downsampling include reducing the time a preview takes to start playing for a user, preventing the preview from starting, stopping, or freezing during playback, and reducing the overall bandwidth usage for the server and the end user.


In one embodiment, a file can be downsampled if it is larger than 360 pixels in the height dimension of the frames. During the downsampling process, the aspect ratio of the frames is maintained. Thus, for sources having a 4:3 aspect ratio, the resulting frame sizes of the video file would be 480×360 pixels, and for wide-format video, the resulting size would be 640×360 pixels. Other downsampling criteria, such as different pixel dimensions of the frames, can also be selected when reducing the size of a file.


In one embodiment, a video file can be downsampled to multiple different conversion sizes, for example, a high resolution version, a medium resolution version, and a low resolution version. In one embodiment, depending upon the tier of service a user pays for, one, two, or all three of the different versions can be made available to the user. For a user paying a premium to subscribe to a higher tier of service, the user would receive more versions than a user subscribing to a lower tier of service. The benefit of receiving more versions having different sizes is that the user then has the option of selecting the most appropriate size or resolution. For example, a low resolution version can be selected for a slow network connection, and a high resolution version can be selected for sharing a video with friends who have a high bandwidth network connection. In one embodiment, if more than one size file is to be produced, the downsampling for each conversion size is performed on the original input file, rather than downsampling a stream of video data multiple times, to minimize the number of conversion errors introduced.


In one embodiment, if a video file is converted from its input format directly to a desired target format, the downsampling can take place prior to the conversion of the file format to reduce the processing time needed to convert the file format. However, if the video file is first converted as part of a multi-step conversion to an intermediate format before being converted to a target format, the unsupported input file format should be converted to the intermediate format first before downsampling for the final conversion to the target format.


With traditional video file downloading, a user is not able to access the video file before it has completely downloaded because a video player needs the metadata encoded in the video file, and the metadata is usually found at the end of the file. Thus, the video player needs to wait until the video file is fully loaded prior to playing the video. One embodiment of the conversion server 400 includes the fast start preview module 450 which can move the metadata from the end of the file to the beginning of the file. By moving the metadata to the beginning of the file, the preview player can begin playing the video before the user has fully downloaded the file. In one embodiment, the fast start preview module 450 uses the qt-faststart library to move the metadata to the beginning of the video file.


One embodiment of the conversion server 400 includes the rotation module 466 which uses a rotation identification tool to identify the embedded orientation data that a digital camera uses to display a video. Once the rotation module 466 has identified the orientation, it sends this information to the conversion tool(s) as a flag to correct the video for playback in the video preview in the online collaboration environment. It is important to identify the orientation data of a video file because video taken upside-down with a digital camera that plays in an appropriate orientation in the digital camera and most desktop video players would appear upside-down in the video preview.



FIG. 5 depicts a flow chart illustrating an example process for determining if a user has permission to access preview of a file.


At block 505, the system receives user input from a user requesting a preview of a file. In one embodiment, the user can request the preview by clicking on the file in the workspace of the online collaboration environment.


Then at block 510, the system determines the permission level of the requesting user. In one embodiment, a first type of permission level, e.g. an editor, allows a user to have full read and write access to a workspace, including downloading or viewing a preview of the contents of the workspace as well as uploading new content to the workspace. A second type of permission level, e.g. a viewer, allows a user to have full read access to a workspace such that the user can download or view a preview of contents of the workspace but not upload or edit contents of the workspace. A third type of permission level, e.g. an uploader, can allow a user to have limited write access to contents of a workspace such that the user can see items in the workspace but not download or view the items, while being permitted to upload new content to the workspace.


At decision block 516, the system determines if the user has permission to preview the file. If the permission level of the user does not allow previewing (block 516—No), at block 526 the system notifies the user that permission to view the preview is denied. If the permission level of the user allows previewing (block 516—Yes), at block 530 the system converts the file to an appropriate format for previewing. Then at block 535 the system presents the preview to the user. In one embodiment, the file is a video file. In this case, the file does not have to finish converting before the file is presented to the user as a preview, as discussed above with respect to streaming a video file.



FIG. 6 depicts a flow chart illustrating an example process for finding an optimum path using different conversion tools for converting an input file format to a target file format.


At block 605, the system identifies the format of an input file that has been uploaded by a user. Then at block 610, the system specifies the desired output format appropriate for presenting a preview of the file. In one embodiment, output size and/or preview type is also specified. Output size can be used to limit the size of the output file. Thus, if the converted file exceeds the specified threshold, the file can be re-processed, reduced in size, or not used. In one embodiment, output dimensions are specified because in different applications, different dimensions of previews may be used. For example, a preview can be displayed as a thumbnail on a file listing page (e.g. FIG. 10) or as a large display on a file preview page (e.g. in FIG. 11). With both of these previews, the files are the same format, but have different dimensions, thus the previews with different dimensions are different system outputs.


Next, at block 615, the system solves the network graph for the optimum conversion scenario using the available conversion tools. The system takes into account the priority cost of using each of the tools to determine the optimum path. Then at block 620, the system converts the file using the identified conversion scenario path.


At decision block 625, the system determines whether the conversion of the input file to the targeted output format was successful. If the conversion was successful (block 625—Yes), the process ends at block 699. If the conversion failed (block 625—No), at block 630, the system identifies the conversion tool that failed to create a sufficient output file. If the faulty conversion tool was used to convert to an intermediate file format, subsequent conversions from the output of the faulty tool would not be successful.


At block 635, the system re-solves the network graph without using the nodes associated with the faulty conversion tool to identify the next optimum path using the available conversion tools. The process returns to block 620 to convert the input file using the newly identified path.



FIG. 7 depicts a flow chart illustrating an example process for converting a format of a vide© file to a target format for previewing.


At block 705, the system begins to receive an uploading video file streamed from a user. Then at decision block 710, the system determines if sufficient data has been received to start converting the file to a desired target format. If sufficient data has not yet been received (block 710—No), at block 715 the system waits and continues to receive the uploading file and returns to decision block 710.


If sufficient data has been received (block 710—Yes), at block 720, the system runs a rotation identification tool on the input file to determine the orientation of the video data. Then at decision block 725, the system determines if the file format is compatible with the conversion tool or tools that the system has available., for example, mencoder. If the file format is not compatible (block 725—No), at block 730 the system converts the input file to an intermediate format that is compatible with the conversion tool and continues to block 735. If the file format is compatible with the conversion tool (block 725—Yes), at block 735 the system specifies the desired output format to the conversion tool.


The system applies the qt-faststart program at block 740 to the file in order to move the metadata to the beginning of the video file. This allows the output file to be streamed to the user while the conversion of the remainder of the file occurs if the user wishes to do so.


Then at block 745, the system uses the video conversion tool to start converting the file, whether it is the input file or an intermediate format file, to the desired format. The conversion process on the file does not have to finish before streaming the converted portions to the user for previewing. The converted portions of the file is stored at block 750. The converted file can be written directly to a publicly-available location in chunks, and the video player can start reading the data and showing the video to the user before the file is finished converting.



FIG. 8 depicts a flow chart illustrating an example process for generating a progress indicator that shows the amount of time remaining for converting a file.


At block 805, the system receives input from a user requesting a file preview. Next, at block 810, the system compares the attributes of the file to analytics stored in a database containing attributes of previously converted files. Attributes can include the size of the file and the file format.


At block 815, the system examines the format of the file to be previewed and determines the number of intermediate steps needed, if any, to convert the file to a target suitable for previewing.


At block 820, the system obtains the current load information on the conversion server(s).


Then at block 825, the system uses the obtained data to determine an expected conversion duration. The system sets a timer and generates a progress indicator for displaying to the user at block 830. The progress of the conversion servers in performing the file conversion can be displayed in any suitable form, such as a bar or a pie.


At block 835, the system checks the timer, and at decision block 840, the system determines if the conversion has finished. If the conversion has completed (block 840—Yes), the process ends at block 899. If the conversion has not finished (block 840—No), at block 845 the system updates the progress indicator and returns to block 835 to check the timer.



FIG. 9 shows a diagrammatic representation of a machine 900 in the example form of a computer system within which a set of instructions, for causing the machine to perform any one or more of the methodologies discussed herein, may be executed.


In alternative embodiments, the machine operates as a standalone device or may be connected (e.g., networked) to other machines. In a networked deployment, the machine may operate in the capacity of a server or a client machine in a client-server network environment, or as a peer machine in a peer-to-peer (or distributed) network environment.


The machine may be a server computer, a client computer, a personal computer (PC), a user device, a tablet PC, a laptop computer, a set-top box (STB), a personal digital assistant (PDA), a cellular telephone, an iPhone, an iPad, a Blackberry, a processor, a telephone, a web appliance, a network router, switch or bridge, a console, a hand-held console, a (hand-held) gaming device, a music player, any portable, mobile, hand-held device, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine.


While the machine-readable medium or machine-readable storage medium is shown in an exemplary embodiment to be a single medium, the term “machine-readable medium” and “machine-readable storage medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “machine-readable medium” and “machine-readable storage medium” shall also be taken to include any medium that is capable of storing, encoding or carrying a set of instructions for execution by the machine and that cause the machine to perform any one or more of the methodologies of the presently disclosed technique and innovation.


In general, the routines executed to implement the embodiments of the disclosure, may be implemented as part of an operating system or a specific application, component, program, object, module or sequence of instructions referred to as “computer programs.” The computer programs typically comprise one or more instructions set at various times in various memory and storage devices in a computer, and that, when read and executed by one or more processing units or processors in a computer, cause the computer to perform operations to execute elements involving the various aspects of the disclosure.


Moreover, while embodiments have been described in the context of fully functioning computers and computer systems, those skilled in the art will appreciate that the various embodiments are capable of being distributed as a program product in a variety of forms, and that the disclosure applies equally regardless of the particular type of machine or computer-readable media used to actually effect the distribution.


Further examples of machine-readable storage media, machine-readable media, or computer-readable (storage) media include, but are not limited to, recordable type media such as volatile and non-volatile memory devices, floppy and other removable disks, hard disk drives, optical disks (e.g., Compact Disk Read-Only Memory (CD ROMS), Digital Versatile Disks, (DVDs), etc.), among others, and transmission type media such as digital and analog communication links.


The network interface device enables the machine 900 to mediate data in a network with an entity that is external to the host server, through any known and/or convenient communications protocol supported by the host and the external entity. The network interface device can include one or more of a network adaptor card, a wireless network interface card, a router, an access point, a wireless router, a switch, a multilayer switch, a protocol converter, a gateway, a bridge, bridge router, a hub, a digital media receiver, and/or a repeater.


The network interface device can include a firewall which can, in some embodiments, govern and/or manage permission to access/proxy data in a computer network, and track varying levels of trust between different machines and/or applications, he firewall can be any number of modules having any combination of hardware and/or software components able to enforce a predetermined set of access rights between a particular set of machines and applications, machines and machines, and/or applications and applications, for example, to regulate the flow of traffic and resource sharing between these varying entities. The firewall may additionally manage and/or have access to an access control list which details permissions including for example, the access and operation rights of an object by an individual, a machine, and/or an application, and the circumstances under which the permission rights stand.


Other network security functions can be performed or included in the functions of the firewall, can be, for example, but are not limited to, intrusion-prevention, intrusion detection, next-generation firewall, personal firewall, etc. without deviating from the novel art of this disclosure.


CONCLUSION

Unless the context clearly requires otherwise, throughout the description and the claims, the words “comprise,” “comprising,” and the like are to be construed in an inclusive sense (i.e., to say, in the sense of “including, but not limited to”), as opposed to an exclusive or exhaustive sense. As used herein, the terms “connected,” “coupled,” or any variant thereof means any connection or coupling, either direct or indirect, between two or more elements. Such a coupling or connection between the elements can be physical, logical, or a combination thereof. Additionally, the words “herein,” “above,” “below,” and words of similar import, when used in this application, refer to this application as a whole and not to any particular portions of this application. Where the context permits, words in the above Detailed Description using the singular or plural number may also include the plural or singular number respectively. The word “or,” in reference to a Hat of two or more items, covers all of the following interpretations of the word: any of the items in the list, all of the items in the list, and any combination of the items in the list.


The above Detailed Description of examples of the invention is not intended to be exhaustive or to limit the invention to the precise form disclosed above. While specific examples for the invention are described above for illustrative purposes, various equivalent modifications are possible within the scope of the invention, as those skilled in the relevant art will recognize. While processes or blocks are presented in a given order in this application, alternative implementations may perform routines having steps performed in a different order, or employ systems having blocks in a different order. Some processes or blocks may be deleted, moved, added, subdivided, combined, and/or modified to provide alternative or subcombinations. Also, while processes or blocks are at times shown as being performed in series, these processes or blocks may instead be performed or implemented in parallel, or may be performed at different times. Further any specific numbers noted herein are only examples. It is understood that alternative implementations may employ differing values or ranges.


The various illustrations and teachings provided herein can also be applied to systems other than the system described above. The elements and acts of the various examples described above can be combined to provide further implementations of the invention.


Any patents and applications and other references noted above, including any that may be listed in accompanying filing papers, are incorporated herein by reference. Aspects of the invention can be modified, if necessary, to employ the systems, functions, and concepts included in such references to provide further implementations of the invention.


These and other changes can be made to the invention in light of the above Detailed Description. While the above description describes certain examples of the invention, and describes the best mode contemplated, no matter how detailed the above appears in text, the invention can be practiced in many ways. Details of the system may vary considerably in its specific implementation, while still being encompassed by the invention disclosed herein. As noted above, particular terminology used when describing certain features or aspects of the invention should not be taken to imply that the terminology is being redefined herein to be restricted to any specific characteristics, features, or aspects of the invention with which that terminology is associated. In general, the terms used in the following claims should not be construed to limit the invention to the specific examples disclosed in the specification, unless the above Detailed Description section explicitly defines such terms. Accordingly, the actual scope of the invention encompasses not only the disclosed examples, but also all equivalent ways of practicing or implementing the invention under the claims.


While certain aspects of the invention are presented below in certain claim forms, the applicant contemplates the various aspects of the invention in any number of claim forms. For example, while only one aspect of the invention is recited as a means-plus-function claim under 35 U.S.C. §112, sixth paragraph, other aspects may likewise be embodied as a means-plus-function claim, or in other forms, such as being embodied in a computer-readable medium. (Any claims intended to be treated under 35 U.S.C. §112, ¶6 will begin with the words “means for.”) Accordingly, the applicant reserves the right to add additional claims after filing the application to pursue such additional claim forms for other aspects of the invention.

Claims
  • 1. A computer implemented method for predicting a duration for converting a file to a target format for previewing using computing elements, the method comprising: comparing attributes of the file with attributes of previously converted files;predicting the duration based upon an average of conversion times for previously converted files with the attributes; andcausing to be displayed an indicator configured to dynamically provide a time remaining for completion of the conversion of the file based upon the predicted duration.
  • 2. The method of claim 1, wherein attributes of the file comprise one or more of: size of the file, format of the file, and number of steps to be used for converting the file, wherein a step corresponds to each format conversion of the file performed to arrive at the target format.
  • 3. The method of claim 2, wherein for text-based files, attributes of the file further comprise a number of pages of the file, wherein for image files, attributes of the file further comprise one or more of a source width and a source height of a frame of the file prior to conversion, and a target width and a target height of a target frame in the target format, and wherein for video files, attributes of the file further comprise one or more of a play length, a source width and a source height of a frame of the file prior to conversion, and a target width and a target height of a target frame in the target format.
  • 4. The method of claim 1, further comprising adjusting the predicted duration based upon a current load on a system used to perform the conversion of the file to the target format.
  • 5. The method of claim 4, further comprising adjusting the predicted duration based upon reported progress of a step of the conversion of the file to the target format.
  • 6. A computer implemented method for decreasing a time for converting a video file to a target format for previewing using computing elements, the method comprising: downsampling the video file to a first downsampled video file, wherein the video file has been uploaded to a web-based collaboration platform;converting the first downsampled video file to a first output file having the target format for previewing using a preview module of a conversion server; andcausing to be displayed the first output file for previewing, wherein the converting is performed on a conversion engine of the conversion server.
  • 7. The method of claim 6, wherein downsampling comprises reducing a height of frames of the video file to a preselected number of pixels while maintaining an aspect ratio of the video file.
  • 8. The method of claim 6, further comprising evaluating a load on a system of distributed conversion servers;using a distributed queuing system to determine which of the distributed conversion servers to use for performing the downsampling and the conversion.
  • 9. The method of claim 8, further comprising using timing information collected from converting previous files to determine which of the distributed conversion servers to use for performing the downsampling and the conversion to increase a throughput of the system.
  • 10. The method of claim 6, further comprising identifying orientation data in the video file and conveying the orientation data to a video player for use in displaying the first output file.
  • 11. The method of claim 6, further comprising: providing a notification of the previewing by the user to another user of the web-based collaboration platform.
  • 12. The method of claim 11, wherein the notification is accessible by the other user when the previewing is displayed.
  • 13. The method of claim 11, wherein the notification is presented in a feed stream through a user interface to the web-based collaboration platform.
  • 14. The method of claim 11, wherein the other user meets a criteria for receiving the notification.
  • 15. A computer implemented method performed using computing elements, the method comprising: downsampling a video file to a first downsampled video file, wherein the video file has been uploaded to a web-based collaboration platform;converting the first downsampled video file to a first output file having a target format; anddownsampling the video file to a second downsampled video file, wherein the second downsampled video file is smaller than the first downsampled video file;converting the second downsampled video file to a second output file having the target format; andproviding the first output file to a user for previewing if the user subscribes to a higher tier of service with the web-based collaboration platform;providing the second output file to the user for previewing if the user subscribes to a lower tier of service with the web-based collaboration platform.
  • 16. The method of claim 15, further comprising: providing the first output file and the second output file to the user to select for previewing, if there is only a single tier of service offered by the web-based collaboration platform.
  • 17. The method of claim 15, wherein downsampling comprises reducing a height of frames of the video file to a preselected number of pixels while maintaining an aspect ratio of the video file.
  • 18. The method of claim 15, further comprising evaluating a load on a system of distributed conversion servers;using a distributed queuing system to determine which of the distributed conversion servers to use for performing the downsampling and the conversion.
  • 19. The method of claim 18, further comprising using timing information collected from converting previous files to determine which of the distributed conversion servers to use for performing the downsampling and the conversion to increase a throughput of the system.
  • 20. The method of claim 15, further comprising: providing a notification of the previewing by the user to another user of the web-based collaboration platform.
  • 21. The method of claim 20, wherein the notification is accessible by the other when the previewing is displayed.
  • 22. The method of claim 20, wherein the notification is presented in a feed stream through a user interface to the web-based collaboration platform.
  • 23. A computer implemented method performed using computing elements, the method comprising: downsampling a video file to a first downsampled video file, wherein the video file has been uploaded to a web-based collaboration platform;converting the first downsampled video file to a first output file having a target format; anddownsampling the video file to a second downsampled video file, wherein the second downsampled video file is smaller than the first downsampled video file;converting the second downsampled video file to a second output file having the target format; andproviding the first output file and the second output file to a user to select for previewing if the user subscribes to a higher tier of service with the web-based collaboration platform;offering only one of the first output file or the second output file to the user for previewing if the user subscribes to a lower tier of service with the web-based collaboration platform.
  • 24. The method of claim 23, wherein downsampling comprises reducing a height of frames of the video file to a preselected number of pixels while maintaining an aspect ratio of the video file.
  • 25. The method of claim 23, further comprising: providing a notification of the previewing by the user to another user of the web-based collaboration platform.
  • 26. The method of claim 25, wherein the notification is accessible by the other user when the previewing is displayed.
  • 27. The method of claim 25, wherein the notification is presented in a feed stream through a user interface to the web-based collaboration platform.
  • 28. A computer implemented method for decreasing a time for converting a video file to a target format for previewing using computing elements, the method comprising: converting the video file to an intermediate format file, wherein the video file has been uploaded to a web-based collaboration platform;downsampling the intermediate format file;converting the downsampled intermediate format file to a target format filecausing to be displayed the target format file for previewing.
  • 29. The method of claim 28, wherein downsampling comprises reducing a height of frames of the video file to a preselected number of pixels while maintaining an aspect ratio of the video file.
  • 30. The method of claim 28, further comprising: providing a notification of the previewing by the user to another user of the web-based collaboration platform.
  • 31. The method of claim 30, wherein the notification is accessible by the other user when the previewing is displayed.
  • 32. The method of claim 30, wherein the notification is presented in a feed stream through a user interface to the web-based collaboration platform.
  • 33. A computer implemented method of streaming a video file to a user for previewing prior to completing conversion of the video file to a target format using computing elements, the method comprising: moving metadata in the video file to a beginning of the video file, wherein the video file has been uploaded to a web-based collaboration platform;as the video file is being converted to the target format, writing sections of the converted portions of the video file to a location accessible by a video player;causing the video player to display the converted portions of the video file to the user while a remainder of the video file is converted to the target format.
  • 34. The method of claim 33, further comprising downsampling the video file prior to converting the video file.
  • 35. The method of claim 33, further comprising evaluating a load on a system of distributed conversion servers;using a distributed queuing system to determine which of the distributed conversion servers to use for performing the conversion.
  • 36. The method of claim 35, further comprising using timing information collected from converting previous files to determine which of the distributed conversion servers to use for performing the conversion to increase a throughput of the system.
  • 37. The method of claim 33, further comprising identifying orientation data in the video file and conveying the orientation data to a video player for use in displaying the first output file.
  • 38. The method of claim 33, further comprising: providing a notification of the previewing by the user to another user of the web-based collaboration platform.
  • 39. The method of claim 38, wherein the notification is accessible by the other user to when the previewing is displayed.
  • 40. The method of claim 38, wherein the notification is presented in a feed stream through a user interface to the web-based collaboration platform.
  • 41. The method of claim 38, wherein the other user meets a criteria for receiving the notification.
  • 42. A system, comprising: one or more conversion servers, wherein each conversion server includes a conversion engine that is configured to convert an input format of a file to a target format; anda progress indicator engine configured to determine a predicted duration for converting the file to the target format and providing a dynamic indicator to indicate to a user a time remaining for the conversion, wherein the target format of the file is used for previewing using a preview module of the one or more conversion servers.
  • 43. The system of claim 42, further comprising a downsampling engine configured to downsample the file to decrease the conversion time.
  • 44. The system of claim 43, further comprising a load balancing engine configured to determine to which of the conversion servers to send the file for conversion.
  • 45. A system, comprising: means for comparing attributes of the file with attributes of previously converted files;means for predicting the duration based upon an average of conversion times for previously converted files with similar attributes; andmeans for causing to be displayed an indicator configured to dynamically provide a time remaining for completion of the conversion of the file based upon the predicted duration.
  • 46. A non-transitory machine readable storage medium encoded with instructions for performing a method for predicting a duration for converting a file to a target format for previewing, the instructions comprising: instructions to compare attributes of the file with attributes of previously converted files;instructions to predict the duration based upon an average of conversion times for previously converted files with the attributes; andinstructions to cause to be displayed, on a display, an indicator configured to dynamically provide a time remaining for completion of the conversion of the file based upon the predicted duration.
  • 47. The non-transitory machine readable storage medium of claim 46, wherein attributes of the file comprise one or more of: size of the file, format of the file, and number of steps to be used for converting the file, wherein a step corresponds to each format conversion of the file performed to arrive at the target format.
  • 48. The non-transitory machine readable storage medium of claim 46, further comprising instructions to adjust the predicted duration based upon a current load on a system used to perform the conversion of the file to the target format.
US Referenced Citations (444)
Number Name Date Kind
5787175 Carter Jul 1998 A
5799320 Klug Aug 1998 A
5848415 Guck Dec 1998 A
5864870 Guck Jan 1999 A
5999908 Abelow Dec 1999 A
6016467 Newsted et al. Jan 2000 A
6034621 Kaufman Mar 2000 A
6055543 Christensen et al. Apr 2000 A
6073161 DeBoskey et al. Jun 2000 A
6098078 Gehani et al. Aug 2000 A
6233600 Salas et al. May 2001 B1
6260040 Kauffman et al. Jul 2001 B1
6289345 Yasue Sep 2001 B1
6292803 Richardson et al. Sep 2001 B1
6336124 Alam et al. Jan 2002 B1
6342906 Kumar et al. Jan 2002 B1
6345386 Delo et al. Feb 2002 B1
6370543 Hoffert et al. Apr 2002 B2
6374260 Hoffert et al. Apr 2002 B1
6385606 Inohara et al. May 2002 B2
6396593 Laverty et al. May 2002 B1
6515681 Knight Feb 2003 B1
6539381 Prasad et al. Mar 2003 B1
6584466 Serbinis et al. Jun 2003 B1
6636872 Heath et al. Oct 2003 B1
6654737 Nunez Nov 2003 B1
6662186 Esquibel et al. Dec 2003 B1
6687878 Eintracht et al. Feb 2004 B1
6714968 Prust Mar 2004 B1
6735623 Prust May 2004 B1
6742181 Koike et al. May 2004 B1
6760721 Chasen et al. Jul 2004 B1
6947162 Rosenberg et al. Sep 2005 B2
6952724 Prust Oct 2005 B2
6996768 Elo et al. Feb 2006 B1
7010752 Ly Mar 2006 B2
7020697 Goodman et al. Mar 2006 B1
7039806 Friedman et al. May 2006 B1
7069393 Miyata et al. Jun 2006 B2
7130831 Howard et al. Oct 2006 B2
7133834 Abelow Nov 2006 B1
7149787 Mutalik et al. Dec 2006 B1
7155483 Friend et al. Dec 2006 B1
7165107 Pouyoul et al. Jan 2007 B2
7222078 Abelow May 2007 B2
7275244 Charles Bell et al. Sep 2007 B1
7296025 Kung et al. Nov 2007 B2
7346778 Guiter et al. Mar 2008 B1
7353252 Yang et al. Apr 2008 B1
7370269 Prabhu et al. May 2008 B1
7386535 Kalucha et al. Jun 2008 B1
7401117 Dan et al. Jul 2008 B2
7543000 Castro et al. Jun 2009 B2
7581221 Lai et al. Aug 2009 B2
7620565 Abelow Nov 2009 B2
7647559 Yozell-Epstein et al. Jan 2010 B2
7661088 Burke Feb 2010 B2
7665093 Maybee et al. Feb 2010 B2
7676542 Moser et al. Mar 2010 B2
7698363 Dan et al. Apr 2010 B2
7734600 Wise et al. Jun 2010 B1
7756843 Palmer Jul 2010 B1
7774412 Schnepel Aug 2010 B1
7814426 Huesken et al. Oct 2010 B2
7886287 Davda Feb 2011 B1
7890964 Vogler-Ivashchanka et al. Feb 2011 B2
7937663 Parker et al. May 2011 B2
7958453 Taing Jun 2011 B1
7979296 Kruse et al. Jul 2011 B2
7996374 Jones et al. Aug 2011 B1
8027976 Ding et al. Sep 2011 B1
RE42904 Stephens, Jr. Nov 2011 E
8065739 Bruening et al. Nov 2011 B1
8090361 Hagan Jan 2012 B2
8103662 Eagan et al. Jan 2012 B2
8117261 Briere et al. Feb 2012 B2
8140513 Ghods et al. Mar 2012 B2
8151183 Chen et al. Apr 2012 B2
8214747 Yankovich et al. Jul 2012 B1
8230348 Peters et al. Jul 2012 B2
8326814 Ghods et al. Dec 2012 B2
8347276 Schadow Jan 2013 B2
8358701 Chou et al. Jan 2013 B2
8370803 Holler et al. Feb 2013 B1
8429540 Yankovich et al. Apr 2013 B1
8464161 Giles et al. Jun 2013 B2
8515902 Savage Aug 2013 B2
8527549 Cidon Sep 2013 B2
8549066 Donahue et al. Oct 2013 B1
8549511 Seki et al. Oct 2013 B2
8583619 Ghods et al. Nov 2013 B2
8607306 Bridge et al. Dec 2013 B1
8719445 Ko May 2014 B2
8745267 Luecke et al. Jun 2014 B2
8849955 Prahlad et al. Sep 2014 B2
8868574 Kiang et al. Oct 2014 B2
8892679 Destagnol et al. Nov 2014 B1
8914900 Smith et al. Dec 2014 B2
20010027492 Gupta Oct 2001 A1
20020029218 Bentley et al. Mar 2002 A1
20020091738 Rohrabaugh et al. Jul 2002 A1
20020099772 Deshpande et al. Jul 2002 A1
20020133509 Johnston et al. Sep 2002 A1
20020147770 Tang Oct 2002 A1
20020194177 Sherman et al. Dec 2002 A1
20030041095 Konda et al. Feb 2003 A1
20030084306 Abburi et al. May 2003 A1
20030093404 Bader et al. May 2003 A1
20030108052 Inoue et al. Jun 2003 A1
20030110264 Whidby et al. Jun 2003 A1
20030115326 Verma et al. Jun 2003 A1
20030135536 Lyons Jul 2003 A1
20030135565 Estrada Jul 2003 A1
20030154306 Perry Aug 2003 A1
20030204490 Kasriel Oct 2003 A1
20030217171 Von Stuermer et al. Nov 2003 A1
20040021686 Barberis Feb 2004 A1
20040088647 Miller et al. May 2004 A1
20040111415 Scardino et al. Jun 2004 A1
20040117438 Considine et al. Jun 2004 A1
20040122949 Zmudzinski et al. Jun 2004 A1
20040128359 Horvitz et al. Jul 2004 A1
20040177138 Salle et al. Sep 2004 A1
20040181579 Huck et al. Sep 2004 A1
20040196307 Zak et al. Oct 2004 A1
20040201604 Kraenzel et al. Oct 2004 A1
20040230624 Frolund et al. Nov 2004 A1
20040246532 Inada Dec 2004 A1
20040267836 Armangau et al. Dec 2004 A1
20050005276 Morgan Jan 2005 A1
20050010860 Weiss et al. Jan 2005 A1
20050050228 Perham et al. Mar 2005 A1
20050063083 Dart et al. Mar 2005 A1
20050097225 Glatt et al. May 2005 A1
20050102328 Ring et al. May 2005 A1
20050108406 Lee et al. May 2005 A1
20050114305 Haynes et al. May 2005 A1
20050114378 Elien et al. May 2005 A1
20050138118 Banatwala et al. Jun 2005 A1
20050182966 Pham et al. Aug 2005 A1
20050198299 Beck et al. Sep 2005 A1
20050198452 Watanabe Sep 2005 A1
20050234864 Shapiro Oct 2005 A1
20050234943 Clarke Oct 2005 A1
20060005163 Huesken et al. Jan 2006 A1
20060026502 Dutta Feb 2006 A1
20060026535 Hotelling et al. Feb 2006 A1
20060036568 Moore et al. Feb 2006 A1
20060041603 Paterson et al. Feb 2006 A1
20060047804 Fredricksen et al. Mar 2006 A1
20060053088 Ali et al. Mar 2006 A1
20060053380 Spataro et al. Mar 2006 A1
20060070083 Brunswig et al. Mar 2006 A1
20060075071 Gillette Apr 2006 A1
20060117247 Fite et al. Jun 2006 A1
20060123062 Bobbitt et al. Jun 2006 A1
20060133340 Rybak et al. Jun 2006 A1
20060168550 Muller et al. Jul 2006 A1
20060174051 Lordi et al. Aug 2006 A1
20060174054 Matsuki Aug 2006 A1
20060179070 George et al. Aug 2006 A1
20060242204 Karas et al. Oct 2006 A1
20060259524 Horton Nov 2006 A1
20060265719 Astl et al. Nov 2006 A1
20060271510 Harward et al. Nov 2006 A1
20070016680 Burd et al. Jan 2007 A1
20070038934 Fellman Feb 2007 A1
20070079242 Jolley et al. Apr 2007 A1
20070100830 Beedubail et al. May 2007 A1
20070115845 Hochwarth et al. May 2007 A1
20070118598 Bedi et al. May 2007 A1
20070124460 McMullen et al. May 2007 A1
20070124737 Wensley et al. May 2007 A1
20070124781 Casey et al. May 2007 A1
20070126635 Houri Jun 2007 A1
20070130143 Zhang et al. Jun 2007 A1
20070130163 Perez et al. Jun 2007 A1
20070198609 Black et al. Aug 2007 A1
20070208878 Barnes-Leon et al. Sep 2007 A1
20070214180 Crawford Sep 2007 A1
20070220016 Estrada et al. Sep 2007 A1
20070220590 Rasmussen et al. Sep 2007 A1
20070240057 Satterfield et al. Oct 2007 A1
20070250762 Mansfield Oct 2007 A1
20070256065 Heishi et al. Nov 2007 A1
20070266304 Fletcher et al. Nov 2007 A1
20070282848 Kiilerich et al. Dec 2007 A1
20070283443 McPherson et al. Dec 2007 A1
20070288290 Motoyama et al. Dec 2007 A1
20080005135 Muthukrishnan et al. Jan 2008 A1
20080005195 Li Jan 2008 A1
20080021959 Naghi et al. Jan 2008 A1
20080028323 Rosen et al. Jan 2008 A1
20080040173 Aleong et al. Feb 2008 A1
20080040503 Kleks et al. Feb 2008 A1
20080046828 Bibliowicz et al. Feb 2008 A1
20080059656 Saliba et al. Mar 2008 A1
20080077631 Petri Mar 2008 A1
20080091763 Devonshire et al. Apr 2008 A1
20080091790 Beck Apr 2008 A1
20080104277 Tian May 2008 A1
20080114720 Smith et al. May 2008 A1
20080133674 Knauerhase et al. Jun 2008 A1
20080140732 Wilson et al. Jun 2008 A1
20080147790 Malaney et al. Jun 2008 A1
20080151817 Fitchett et al. Jun 2008 A1
20080182628 Lee et al. Jul 2008 A1
20080183467 Yuan et al. Jul 2008 A1
20080184130 Tien et al. Jul 2008 A1
20080194239 Hagan Aug 2008 A1
20080215883 Fok et al. Sep 2008 A1
20080222654 Xu et al. Sep 2008 A1
20080243855 Prahlad et al. Oct 2008 A1
20080250333 Reeves et al. Oct 2008 A1
20080250348 Alimpich et al. Oct 2008 A1
20080263099 Brady-Kalnay et al. Oct 2008 A1
20080271095 Shafton Oct 2008 A1
20080276158 Lim et al. Nov 2008 A1
20080294899 Gazzetta et al. Nov 2008 A1
20090015864 Hasegawa Jan 2009 A1
20090019093 Brodersen et al. Jan 2009 A1
20090019426 Baeumer et al. Jan 2009 A1
20090030710 Levine Jan 2009 A1
20090044128 Baumgarten et al. Feb 2009 A1
20090049131 Lyle et al. Feb 2009 A1
20090119322 Mills et al. May 2009 A1
20090125469 McDonald et al. May 2009 A1
20090132651 Roger et al. May 2009 A1
20090138808 Moromisato et al. May 2009 A1
20090150417 Ghods et al. Jun 2009 A1
20090158142 Arthursson et al. Jun 2009 A1
20090164438 Delacruz Jun 2009 A1
20090171983 Samji et al. Jul 2009 A1
20090193107 Srinivasan et al. Jul 2009 A1
20090193345 Wensley et al. Jul 2009 A1
20090198772 Kim et al. Aug 2009 A1
20090210459 Nair et al. Aug 2009 A1
20090214115 Kimura et al. Aug 2009 A1
20090235167 Boyer et al. Sep 2009 A1
20090235181 Saliba et al. Sep 2009 A1
20090235189 Aybes et al. Sep 2009 A1
20090249224 Davis et al. Oct 2009 A1
20090254589 Nair et al. Oct 2009 A1
20090260060 Smith et al. Oct 2009 A1
20090265430 Bechtel et al. Oct 2009 A1
20090271708 Peters et al. Oct 2009 A1
20090276771 Nickolov et al. Nov 2009 A1
20090282212 Peterson Nov 2009 A1
20090300527 Malcolm et al. Dec 2009 A1
20090327358 Lukiyanov et al. Dec 2009 A1
20090327961 De Vorchik et al. Dec 2009 A1
20100011292 Marinkovich et al. Jan 2010 A1
20100011447 Jothimani Jan 2010 A1
20100017262 Iyer et al. Jan 2010 A1
20100036929 Scherpa et al. Feb 2010 A1
20100042720 Stienhans et al. Feb 2010 A1
20100057560 Skudlark et al. Mar 2010 A1
20100057785 Khosravy et al. Mar 2010 A1
20100076946 Barker et al. Mar 2010 A1
20100082634 Leban Apr 2010 A1
20100083136 Komine et al. Apr 2010 A1
20100088150 Mazhar et al. Apr 2010 A1
20100092126 Kaliszek et al. Apr 2010 A1
20100093310 Gbadegesin et al. Apr 2010 A1
20100107225 Spencer et al. Apr 2010 A1
20100131868 Chawla et al. May 2010 A1
20100151431 Miller Jun 2010 A1
20100153835 Xiong et al. Jun 2010 A1
20100162365 Del Real Jun 2010 A1
20100162374 Nair Jun 2010 A1
20100185463 Noland et al. Jul 2010 A1
20100185932 Coffman et al. Jul 2010 A1
20100191689 Cortes et al. Jul 2010 A1
20100198783 Wang et al. Aug 2010 A1
20100198871 Stiegler et al. Aug 2010 A1
20100198944 Ho et al. Aug 2010 A1
20100205537 Knighton et al. Aug 2010 A1
20100218237 Ferris et al. Aug 2010 A1
20100223378 Wei Sep 2010 A1
20100229085 Nelson et al. Sep 2010 A1
20100235526 Carter et al. Sep 2010 A1
20100235539 Carter et al. Sep 2010 A1
20100241611 Zuber Sep 2010 A1
20100241972 Spataro et al. Sep 2010 A1
20100250120 Waupotitsch et al. Sep 2010 A1
20100251340 Martin et al. Sep 2010 A1
20100257457 De Goes Oct 2010 A1
20100262582 Garcia-Ascanio et al. Oct 2010 A1
20100267588 Nelson et al. Oct 2010 A1
20100274765 Murphy et al. Oct 2010 A1
20100274772 Samuels Oct 2010 A1
20100281118 Donahue et al. Nov 2010 A1
20100306379 Ferris Dec 2010 A1
20100318893 Matthews et al. Dec 2010 A1
20100322252 Suganthi et al. Dec 2010 A1
20100325155 Skinner et al. Dec 2010 A1
20100325527 Estrada et al. Dec 2010 A1
20100325559 Westerinen et al. Dec 2010 A1
20100325655 Perez Dec 2010 A1
20100332401 Prahlad et al. Dec 2010 A1
20100332962 Hammer et al. Dec 2010 A1
20100333116 Prahlad et al. Dec 2010 A1
20110001763 Murakami Jan 2011 A1
20110016409 Grosz et al. Jan 2011 A1
20110022559 Andersen et al. Jan 2011 A1
20110029883 Lussier et al. Feb 2011 A1
20110040812 Phillips Feb 2011 A1
20110041083 Gabai et al. Feb 2011 A1
20110047413 McGill et al. Feb 2011 A1
20110047484 Mount et al. Feb 2011 A1
20110052155 Desmarais et al. Mar 2011 A1
20110054968 Galaviz Mar 2011 A1
20110055299 Phillips Mar 2011 A1
20110055721 Jain et al. Mar 2011 A1
20110061045 Phillips Mar 2011 A1
20110061046 Phillips Mar 2011 A1
20110065082 Gal et al. Mar 2011 A1
20110066951 Ward-Karet et al. Mar 2011 A1
20110093567 Jeon et al. Apr 2011 A1
20110099006 Sundararaman et al. Apr 2011 A1
20110107205 Chow et al. May 2011 A1
20110113320 Neff et al. May 2011 A1
20110119313 Sung et al. May 2011 A1
20110137991 Russell Jun 2011 A1
20110142410 Ishii Jun 2011 A1
20110145744 Haynes et al. Jun 2011 A1
20110161289 Pei et al. Jun 2011 A1
20110167125 Achlioptas Jul 2011 A1
20110167353 Grosz et al. Jul 2011 A1
20110185292 Chawla et al. Jul 2011 A1
20110202424 Chun et al. Aug 2011 A1
20110202599 Yuan et al. Aug 2011 A1
20110209064 Jorgensen et al. Aug 2011 A1
20110219419 Reisman Sep 2011 A1
20110225417 Maharajh et al. Sep 2011 A1
20110238458 Purcell et al. Sep 2011 A1
20110238621 Agrawal Sep 2011 A1
20110239135 Spataro et al. Sep 2011 A1
20110246294 Robb et al. Oct 2011 A1
20110246950 Luna et al. Oct 2011 A1
20110252071 Cidon Oct 2011 A1
20110252320 Arrasvuori et al. Oct 2011 A1
20110252339 Lemonik et al. Oct 2011 A1
20110258461 Bates Oct 2011 A1
20110258561 Ladouceur et al. Oct 2011 A1
20110289433 Whalin et al. Nov 2011 A1
20110313803 Friend et al. Dec 2011 A1
20110320197 Conejero et al. Dec 2011 A1
20120036370 Lim et al. Feb 2012 A1
20120064879 Panei Mar 2012 A1
20120072436 Pierre et al. Mar 2012 A1
20120079095 Evans et al. Mar 2012 A1
20120089659 Halevi et al. Apr 2012 A1
20120092055 Peschke et al. Apr 2012 A1
20120110005 Kuo et al. May 2012 A1
20120110436 Adler, III et al. May 2012 A1
20120110443 Lemonik et al. May 2012 A1
20120117626 Yates et al. May 2012 A1
20120124306 Abercrombie et al. May 2012 A1
20120124547 Halbedel May 2012 A1
20120130900 Tang et al. May 2012 A1
20120134491 Liu May 2012 A1
20120136936 Quintuna May 2012 A1
20120144283 Hill et al. Jun 2012 A1
20120150888 Hyatt et al. Jun 2012 A1
20120151551 Readshaw et al. Jun 2012 A1
20120158908 Luna et al. Jun 2012 A1
20120159178 Lin et al. Jun 2012 A1
20120159310 Chang et al. Jun 2012 A1
20120173625 Berger Jul 2012 A1
20120179981 Whalin et al. Jul 2012 A1
20120185355 Kilroy Jul 2012 A1
20120192055 Antebi et al. Jul 2012 A1
20120192086 Ghods et al. Jul 2012 A1
20120204032 Wilkins et al. Aug 2012 A1
20120214444 McBride et al. Aug 2012 A1
20120218885 Abel et al. Aug 2012 A1
20120221789 Felter Aug 2012 A1
20120226767 Luna et al. Sep 2012 A1
20120233155 Gallmeier et al. Sep 2012 A1
20120233205 McDermott Sep 2012 A1
20120233543 Vagell et al. Sep 2012 A1
20120240061 Hillenius et al. Sep 2012 A1
20120257249 Natarajan Oct 2012 A1
20120263166 Cho et al. Oct 2012 A1
20120266203 Elhadad et al. Oct 2012 A1
20120284638 Cutler et al. Nov 2012 A1
20120284664 Zhao Nov 2012 A1
20120296790 Robb Nov 2012 A1
20120309540 Holme et al. Dec 2012 A1
20120311157 Erickson et al. Dec 2012 A1
20120317239 Mulder et al. Dec 2012 A1
20120317487 Lieb et al. Dec 2012 A1
20120328259 Seibert, Jr. et al. Dec 2012 A1
20120331177 Jensen Dec 2012 A1
20120331441 Adamson Dec 2012 A1
20130007245 Malik et al. Jan 2013 A1
20130007471 Grab et al. Jan 2013 A1
20130007894 Dang et al. Jan 2013 A1
20130013560 Goldberg et al. Jan 2013 A1
20130014023 Lee et al. Jan 2013 A1
20130042106 Persaud et al. Feb 2013 A1
20130055127 Saito et al. Feb 2013 A1
20130067232 Cheung et al. Mar 2013 A1
20130073403 Tuchman et al. Mar 2013 A1
20130080919 Kiang et al. Mar 2013 A1
20130117337 Dunham May 2013 A1
20130117376 Filman et al. May 2013 A1
20130124638 Barreto et al. May 2013 A1
20130138608 Smith May 2013 A1
20130138615 Gupta et al. May 2013 A1
20130159411 Bowen Jun 2013 A1
20130163289 Kim et al. Jun 2013 A1
20130167253 Seleznev et al. Jun 2013 A1
20130185347 Romano Jul 2013 A1
20130185558 Seibert et al. Jul 2013 A1
20130191339 Haden et al. Jul 2013 A1
20130198600 Lockhart et al. Aug 2013 A1
20130212486 Joshi et al. Aug 2013 A1
20130218978 Weinstein et al. Aug 2013 A1
20130239049 Perrodin et al. Sep 2013 A1
20130246932 Zaveri et al. Sep 2013 A1
20130262210 Savage et al. Oct 2013 A1
20130262862 Hartley Oct 2013 A1
20130268480 Dorman Oct 2013 A1
20130268491 Chung et al. Oct 2013 A1
20130275398 Dorman et al. Oct 2013 A1
20130275429 York et al. Oct 2013 A1
20130275509 Micucci et al. Oct 2013 A1
20130282830 Besen et al. Oct 2013 A1
20130305039 Gauda Nov 2013 A1
20130326344 Masselle et al. Dec 2013 A1
20140013112 Cidon et al. Jan 2014 A1
20140019497 Cidon et al. Jan 2014 A1
20140019498 Cidon et al. Jan 2014 A1
20140032489 Hebbar et al. Jan 2014 A1
20140032616 Nack Jan 2014 A1
20140033277 Xiao et al. Jan 2014 A1
20140033291 Liu Jan 2014 A1
20140052939 Tseng et al. Feb 2014 A1
20140068589 Barak Mar 2014 A1
20140150023 Gudorf et al. May 2014 A1
20140156373 Roberts et al. Jun 2014 A1
20140172595 Beddow et al. Jun 2014 A1
Foreign Referenced Citations (43)
Number Date Country
2724521 Nov 2009 CA
101997924 Mar 2011 CN
102264063 Nov 2011 CN
0348614 Jan 1990 EP
0921661 Jun 1999 EP
1349088 Oct 2003 EP
1528746 May 2005 EP
1933242 Jun 2008 EP
2372574 Oct 2011 EP
2610776 Jul 2013 EP
2453924 Apr 2009 GB
2471282 Dec 2010 GB
09-101937 Apr 1997 JP
11-025059 Jan 1999 JP
2003273912 Sep 2003 JP
2004310272 Nov 2004 JP
09-269925 Oct 2007 JP
2008250944 Oct 2008 JP
20020017444 Mar 2002 KR
20040028036 Apr 2004 KR
20050017674 Feb 2005 KR
20060070306 Jun 2006 KR
20060114871 Nov 2006 KR
20070043353 Apr 2007 KR
20070100477 Oct 2007 KR
20100118836 Nov 2010 KR
20110074096 Jun 2011 KR
20110076831 Jul 2011 KR
WO-0007104 Feb 2000 WO
WO-0219128 Mar 2002 WO
WO-2004097681 Nov 2004 WO
WO-2006028850 Mar 2006 WO
WO-2007024438 Mar 2007 WO
WO-2007035637 Mar 2007 WO
WO-2007113573 Oct 2007 WO
WO-2008011142 Jan 2008 WO
WO-2008076520 Jun 2008 WO
WO-2011109416 Sep 2011 WO
WO-2012167272 Dec 2012 WO
WO-2013009328 Jan 2013 WO
WO-2013013217 Jan 2013 WO
WO-2013041763 Mar 2013 WO
WO-2013166520 Nov 2013 WO
Non-Patent Literature Citations (116)
Entry
Exam Report for EP13168784.0, Applicant: Box, Inc. Mailed Nov. 21, 2013, 7 pages.
Exam Report for EP13185269.1, Applicant: Box, Inc. Mailed Jan. 28, 7 pages.
Exam Report for GB1311417.8, Applicant: Box, Inc. Mailed Dec. 20, 2013, 5 pages.
Exam Report for GB1312095.1, Applicant: Box, Inc. Mailed Dec. 12, 2013, 7 pages.
Exam Report for GB1312874.9, Applicant: Box, Inc. Mailed Dec. 20, 2013, 11 pages.
Exam Report for GB1316971.9, Applicant: Box, Inc. Mailed Nov. 26, 2013, 10 pages.
Exam Report for GB1317600.3, Applicant: Box, Inc. Mailed Nov. 21, 2013, 8 pages.
Exam Report for GB1318373.6, Applicant: Box, Inc. Mailed Dec. 17, 2013, 4 pages.
Exam Report for GB1320902.8, Applicant: Box, Inc. Mailed Dec. 20, 2013, 4 pages.
International Search Report and Written Opinion for PCT/US2013/034765, Applicant: Box, Inc., Mailed Jan. 20, 2014, 15 pages.
Patent Court Document of Approved Judgment for GB0602349.3 and GB0623571.7; Mar. 3, 2009, 17 pages.
Parr, “Google Docs Improves Commenting, Adds E-mail Notifications,” Apr. 16, 2011, mashable.com, pp. 1-6.
“Microsoft Office SharePoint 2007 User Guide,” Feb. 16, 2010, pp. 1-48.
Lars, “35 Very Useful Online Tools for Improving your project Management and Team Collaboration,” Apr. 31, 2010, tripwiremagazine.com, pp. 1-32.
International Search Report and Written Opinion for PCT/US2010/070366, Applicant: Box, Inc., Mailed Mar. 24, 2013, 10 pages.
“How-to Geek, How to Sync Specific Folders With Dropbox,” downloaded from the internet http://www.howtogeek.com, Apr. 23, 2013, 5 pages.
International Search Report and Written Opinion for PCT/US2013/020267, Applicant: Box, Inc., Mailed May 7, 2013, 10 pages.
International Search Report and Written Opinion for PCT/US2013/034662, Applicant: Box, Inc., Mailed May 31, 2013, 10 pages.
“PaperPort Professional 14,” PC Mag. Com review, published Feb. 2012, Ziff Davis, Inc., 8 pages.
“PaperPort,” Wikipedia article (old revision), published May 19, 2012, Wikipedia Foundation, 2 pages.
“Quickoffice Enhances Android Mobile office Application for Improved Productivity on latest Smartphone and Table Devices,” QuickOffice Press Release, Nov. 21, 2011, QuickOffice Inc., 2 pages.
“QuickOffice,” Wikipedia Article (old revision), published May 9, 2012, Wikipedia Foundation, 2 pages.
Exam Report for GB1309209.3, Applicant: Box, Inc. Mailed Oct. 30, 2013, 11 pages.
Exam Report for GB1310666.1, Applicant: Box, Inc. Mailed Aug. 30, 2013, 10 pages.
Exam Report for GB1313559.5, Applicant: Box, Inc., Mailed Aug. 22, 2013, 19 pages.
Exam Report for GB1316532.9, Applicant: Box, Inc. Mailed Oct. 31, 2013, 10 pages.
Exam Report for GB1316533.7, Applicant: Box, Inc. Mailed Oct. 8, 2013, 9 pages.
Gedymin, “Cloud computing with an emphasis on Google App Engine,” Master Final Project, Sep. 2011, 146 pages.
Google Docs, http://web.Archive.org/web/20100413105758/http://en.wikipedia.org/wiki/Google—docs, Apr. 13, 2010, 6 pages.
International Search Report and Written Opinion for PCT/US2013/039782, Applicant: Box, Inc., Mailed Aug. 28, 2013, 15 pages.
“Understanding Metadata,” National Information Standards Organization, NISO Press, 2004, 20 pages.
International Search Report and Written Opinion for PCT/US2012/056955, Applicant: Box, Inc., Mailed Mar. 27, 2013, pp. 1-11.
International Search Report and Written Opinion for PCT/US2011/047530, Applicant: Box, Inc., Mailed Mar. 22, 2013, pp. 1-10.
International Search Report and Written Opinion for PCT/US2012/065617, Applicant: Box, Inc., Mailed Mar. 29, 2013, 9 pages.
International Search Report and Written Opinion for PCT/US2012/067126, Applicant: Box, Inc., Mailed Mar. 29, 2013, 10 pages.
International Search Report and Written Opinion for PCT/US2012/063041, Applicant: Box, Inc., Mailed Mar. 29, 2013, 12 pages.
International Search Report and Written Opinion for PCT/US2011/057938, Applicant: Box, Inc., Mailed Mar. 29, 2013, 10 pages.
“Conceptboard”, One-Step Solution for Online Collaboration, retrieved from websites http://conceptboard.com and https://www.youtube.com/user/ConceptboardApp?feature=watch, printed on Jun. 13, 2013, 9 pages.
Exam Report for EP13158415.3, Applicant: Box, Inc. Mailed Jun. 4, 2013, 8 pages.
Exam Report for GB1300188.8, Applicant: Box, Inc. Mailed May 31, 2013, 8 pages.
Exam Report for GB1306011.6, Applicant: Box, Inc. Mailed Apr. 18, 2013, 8 pages.
International Search Report and Written Opinion for PCT/US2013/023889, Applicant: Box, Inc., Mailed Jun. 24, 2013, 13 pages.
International Search Report and Written Opinion for PCT/US2013/029520, Applicant: Box, Inc., Mailed Jun. 26, 2013, 10 pages.
International Search Report and Written Opinion for PCT/US2013/035404, Applicant: Box, Inc., Mailed Jun. 26, 2013, 13 pages.
International Search Report and Written Opinion for PCT/US2011/039126 mailed on Oct. 6, 2011, pp. 1-13.
Partial International Search Report for PCT/US2011/041308 dated Feb. 27, 2012, pp. 1-2.
International Search Report and Written Opinion for PCT/US2011/056472 mailed on Jun. 22, 2012, pp. 1-12.
Langfeld L. et al., “Microsoft SharePoint 2003 Unleashed,” Chapters 11 and 15, Jun. 2004, pp. 403-404, 557-561, 578-581.
International Search Report and Written Opinion for PCT/US2011/041308 Mailed Jul. 2, 2012, pp. 1-16.
U.S. Appl. No. 13/030,090, filed Feb. 17, 2011, Ghods et al.
U.S. Appl. No. 13/152,982, filed Jun. 3, 2011, Ghods et al.
U.S. Appl. No. 13/165,725, filed Jun. 21, 2011, Trombley-Shapiro et al.
U.S. Appl. No. 13/208,615, filed Aug. 12, 2011, Lee et al.
International Search Report PCT/US2008/012973 dated Apr. 30, 2009, pp. 1-3.
Supplementary European Search Report European Application No. EP 08 85 8563 dated Jun. 20, 2011 pp. 1-5.
Written Opinion PCT/US2008/012973 dated Apr. 30, 2009, pp. 1-4.
International Search Report and Written Opinion for PCT/US2011/060875 Mailed Oct. 30, 2012, pp. 1-10.
“Revolving sync conflicts; frequently asked questions,” Microsoft Tech Support, Jul. 16, 2012, retrieved from the Internet: http://web.archive.org/web, 2 pages.
“Troubleshoot sync problems,” Microsoft Tech Support: May 2, 2012, retrieved from the internet, http://web. Archive.org/web, 3 pages.
“Tulsa TechFest 2012—Agenda,” retrieved from the website, http://web.archive.org, Oct. 2, 2012, 2 pages.
Cohen, “Debating the Definition of Cloud Computing Platforms,” retrieved from the internet, http://forbes.com, Feb. 3, 2014, 7 pages.
Delendik, “Evolving with Web Standards—The Story of PDF.JS,” retrieved from the internet, http://people.mozilla.org, Oct. 12, 2012, 36 pages.
Delendik, “My PDF.js talk slides from Tulsa TechFest,” retrieved from the internet, http://twitter.com, Oct. 12, 2012, 2 pages.
Duffy, “The Best File-Syncing Services,” pcmag.com, retrieved from the internet: http://www.pcmag.com, Sep. 28, 2012, 7 pages.
Exam Report for EP13177108.1, Applicant: Box, Inc. Mailed May 26, 2014, 6 pages.
Exam Report for GB1312264.3, Applicant: Box, Inc. Mailed Mar. 24, 2014, 7 pages.
Exam Report for GB1318792.7, Applicant: Box, Inc. Mailed May 22, 2014, 2 pages.
John et al., “Always Sync Support Forums—View topic—Allway sync funny behavior,” Allway Sync Support Forum at http://sync-center.com, Mar. 28, 2011, XP055109680, 2 pages.
Partial Search Report for EP131832800, Applicant: Box, Inc. Mailed May 8, 2014, 5 pages.
Pyle et al., “How to enable Event logging for Offline Files (Client Side Caching) in Windows Vista,” Feb. 18, 2009, retrieved from the internet: http://blogs.technet.com, 3 pages.
Rao, “Box Acquires Crocodoc To Add HTML5 Document Converter And Sleek Content Viewing Experience To Cloud Storage Platform,” retrieved from the internet, http://techcrunch.com, May 9, 2013, 8 pages.
Search Report for EP13187217.8, Applicant: Box, Inc. Mailed Apr. 15, 2014, 12 pages.
Search Report for EP141509422, Applicant: Box, Inc. Mailed May 8, 2014, 7 pages.
Search Report for EP14151588.2, Applicant: Box, Inc. Mailed Apr. 15, 2014, 12 pages.
Search Report for EP14153783.7, Applicant: Box, Inc. Mailed Apr. 3, 2014, 6 pages.
Sommerer, “Presentable Document Format: Improved On-demand PDF to HTML Conversion,” retrieved from the internet, http://research.microsoft.com, 8 pages.
Tulloch et al., “Windows Vista Resource Kit,” Apr. 8, 2007, Microsoft Press, XP055113067, 6 pages.
Walker, “PDF.js project meeting notes,” retrieved from the internet, http://groups.google.com, May 15, 2014, 1 page.
Exam Report for GB1314771.5, Applicant: Box, Inc. Mailed Feb. 17, 2014, 7 pages.
Exam Report for GB1308842.2, Applicant: Box, Inc. Mailed Mar. 10, 2014, 4 pages.
Burns, “Developing Secure Mobile Applications for Android,” Oct. 2008, Version 1.0, 1-28 pages.
Search Report for EP 11729851.3, Applicant: Box, Inc. Mailed Feb. 7, 2014, 9 pages.
Comes, “MediaXchange User's Manual,” Version 1.15.15, Feb. 1, 2009, pp. 1-90.
“Average Conversion Time for a D60 RAW file?” http://www.dpreview.com, Jul. 22, 2002, 4 pages.
Exam Report for GB1410569.6 Applicant: Box, Inc. Mailed Jul. 11, 2014, 9 pages.
Sommerer, “Presentable Document Format: Improved On-demand PDF to HTML Conversion,” retrieved from the internet, http://research.microsoft.com, Nov. 2004, 8 pages.
Exam Report for GB1316682.2 Applicant: Box, Inc. Mailed Nov. 19, 2014, 6 pages.
Exam Report for GB1312095.1 Applicant: Box, Inc. Mailed Nov. 19, 2014, 5 pages.
Exam Report for GB1313559.5 Applicant: Box, Inc. Mailed Nov. 4, 2014, 2 pages.
User's Guide for SMART Board Software for Windows, published Dec. 2004, 90 pages.
Zambonini et al., “Automated Measuring of Interaction with User Interfaces,” Published as WO2007113573 Oct. 2007, 19 pages.
Exam Report for GB1312874.9 Applicant: Box, Inc. Mailed Sep. 26, 2014, 2 pages.
Exam Report for GB1415126.0 Applicant: Box, Inc. Mailed Oct. 2, 2014, 8 pages.
Exam Report for GB1415314.2 Applicant: Box, Inc. Mailed Oct. 7, 2014, 6 pages.
Exam Report for GB1309209.3 Applicant: Box, Inc. Mailed Oct. 7, 2014, 3 pages.
Exam Report for GB1315232.7 Applicant: Box, Inc. Mailed Oct. 9, 2014, 5 pages.
Exam Report for GB1318789.3 Applicant: Box, Inc. Mailed Oct. 30, 2014, 6 pages.
Microsoft Windows XP Professional Product Documentation: How Inheritance Affects File and Folder Permissions, Apr. 11, 2014, 2 pages.
Exam Report for GB1317393.5 Applicant: Box, Inc. Mailed Nov. 7, 2014, 6 pages.
Exam Report for GB1311417.8 Applicant: Box, Inc. Mailed Nov. 7, 2014, 2 pages.
Exam Report for GB1311421.0 Applicant: Box, Inc. Mailed Nov. 7, 2014, 4 pages.
Internet Forums, http://web.archive.org/web/20100528195550/http://en.wikipedia.org/wiki/Internet—forums, Wikipedia, May 30, 2010, pp. 1-20.
Yahoo! Groups, http://web.archive.org/web/20090320101529/http://en.wikipedia.org/wiki/Yahoo!—Groups, Wikipedia, Mar. 20, 2009, pp. 1-6.
Wiki, http://web.archive.org/web/20100213004936/http://en.wikipedia.org/wiki/Wiki, Feb. 13, 2010, pp. 1-16.
Conner, “Google Apps: The Missing Manual,” published by O'Reilly Media, May 27, 2008, 24 pages.
Cisco, “FTP Load Balancing on ACE in Routed Mode Configuration Example,” DocWiki, Jun. 2011, 7 pages.
Palmer, “Load Balancing FTP Servers,” BlogNav, Oct. 2008, 2 pages.
Wayback, “Wayback machine,” Wayback, Jun. 1, 2011, 1 page.
“Agilewords—How to Request Approval,” YouTube, http://www.youtube.com/watch?v=3Ov3DYNN3Q, Jan. 31, 2011, 2 pages.
“Agilewords—Features, Powerful Features Yet Simple,” Jun. 1, 2011, http://web.archive.org/web/20110601223756/http://agilewords.com/product/features, 3 pages.
Conner, “Google Apps: The Missing Manual,” published by O'Reilly Media, May 27, 2008, 42 pages.
Exam Report for EP 13177108.1, Applicant: Box, Inc. Mailed Feb. 17, 2015, 6 pages.
Exam Report for GB1312264.3 Applicant: Box, Inc. Mailed Jan. 30, 2015, 5 pages.
Exam Report for GB1312874.9 Applicant: Box, Inc. Mailed Feb. 10, 2015, 7 pages.
Exam Report for GB1316685.5 Applicant: Box, Inc. Mailed Feb. 17, 2015, 5 pages.
Exam Report for EP 13185269.1, Applicant: Box, Inc. Mailed Feb. 13, 2015, 8 pages.
Related Publications (1)
Number Date Country
20120328259 A1 Dec 2012 US