The present invention relates generally to computer security.
Social networks, such as the TWITTER, FACEBOOK, and REDDIT social networks, allow end-users to post and share messages. Social networks have become so widespread that end-users access them not only during personal time but also during work hours. In particular, employees may use social networks for productivity, work, personal, or other reasons. One problem with employee access to social networks is that the content exchanged on these networks is very broad and may be against company policy. It is thus desirable for companies and other organizations to be able to control the social network usage on their computer network.
In one embodiment, social network usage in an enterprise environment is controlled by receiving and processing dynamic postings from a social network to identify indicators of prohibited content. The indicators of prohibited content are employed to identify and block prohibited postings from entering an enterprise network.
These and other features of the present invention will be readily apparent to persons of ordinary skill in the art upon reading the entirety of this disclosure, which includes the accompanying drawings and claims.
The use of the same reference label in different drawings indicates the same or like components.
In the present disclosure, numerous specific details are provided, such as examples of apparatus, components, and methods, to provide a thorough understanding of embodiments of the invention. Persons of ordinary skill in the art will recognize, however, that the invention can be practiced without one or more of the specific details. In other instances, well-known details are not shown or described to avoid obscuring aspects of the invention.
Social network usage in enterprise environments has become so prevalent that many computer security products provide a solution for controlling social network usage. Generally speaking, these solutions consist of identifying applications involving social networks and then either blocking the applications or monitoring them. Other solutions that claim to monitor so-called “Shadow IT” go to the next level by doing log analysis along with packet analysis to determine the application usage and the volume of usage per user basis, etc. These solutions claim to have advanced features that include the ability to control certain aspects of the social network applications in a granular way, such as prohibiting posting of pictures but allow viewing.
A major problem with currently available solutions for controlling social network usage in an enterprise environment is their inability to effectively analyze dynamic content, i.e., new content that has just been created. For example, the TWITTER and FACEBOOK social networks have a so-called “trending” feature. In a nutshell, trending comprises updates or tweets that are specifically tagged for interest counting (e.g., user clicks, viewing) and, when the count reaches a particular threshold, the social network starts showing them as “trending” topics. The YAHOO! web portal has similar features. The contents of these trending topics are dynamic in that the contents vary, such as every minute or even more frequently. Because of the dynamic nature of the trending topics, it is relatively difficult for an enterprise to identify the type of content included in the trending topics. More particularly, unlike a static webpage that stays unchanged long enough for a computer security vendor to detect and categorize, dynamic contents are constantly changing as end-users add to them, making categorization relatively difficult. As a particular example, assuming an enterprise's security policy does not allow any “riots” related content to be on their computer network, the enterprise may not be able to block updates or tweets that are tagged “Ferguson” even though there was a riot in Ferguson, Mo.
Referring now to
The computer 100 may be configured to perform its functions by executing the software modules 110. The software modules 110 may be loaded from the data storage device 106 to the main memory 108. An article of manufacture may be embodied as computer-readable storage medium including instructions that when executed by the computer 100 causes the computer 100 to be operable to perform the functions of the software modules 110.
A social network 230 may comprise one or more computers that are configured to provide a social network service. A social network 230 may comprise the social network infrastructure of the TWITTER, REDDIT, FACEBOOK, or other social network. End-users of a social network create messages by posting. A posting may be a tweet in the case of the TWITTER social network, a thread (also referred to as a “reddit”) in the case of the REDDIT social network, or simply a post in the case of the FACEBOOK social network. A posting may also be other user-created messages on other social networks.
In the example of
The content enforcement endpoint 241 and/or the security system 220 may be on-premise within the perimeter of enterprise network 240. The content enforcement endpoint 241 and/or the security system 220 may also be off-premise (i.e., outside the perimeter of the enterprise network 240), such as in the case when the control of social network usage is provided in-the-cloud by a third-party computer security service (e.g., TREND MICRO SMART PROTECTION NETWORK security service).
In the example of
The security system 220 may comprise one or more computers that are configured to receive dynamic postings from a social network 230, to set filtering criteria for filtering a posting stream from the social network 230, to receive and index filtered postings that are filtered from the posting stream based on the filtering criteria, to process the filtered postings to identify prohibited postings, and to provide the content enforcement endpoint 241 the identity and/or indicators of the prohibited postings. The security system 220 may be operated by the administrator or other information technology (IT) personnel of the enterprise network 240 or computer security vendor using scripts, software development kits (SDK), application programming interface (API), and other tools.
In an example operation, a social network 230-1 may be providing a posting stream to one or more computers of the enterprise network 240 (arrow 201). The posting stream may include permitted and prohibited postings. That is, the posting stream as received by the content enforcement endpoint 241 may include postings that are prohibited by one or more computer security policies of the enterprise network 240. The posting stream may also include postings that are not specifically prohibited by a computer security policy, and are thus permitted on the enterprise network 240.
In the example of
The security system 220 may be configured to ingest and process dynamic postings to identify prohibited postings or characteristics of prohibited postings. In one embodiment, uniform resource locators (URLs) extracted from dynamic postings, sites that are linked in the dynamic postings, etc. may be categorized by, for example, consulting a web reputation system 244 (arrow 203). The web reputation system 244 may indicate the category of the URL (e.g., whether the URL or site linked by the URL belongs to the “sports”, “politics”, “pornography,” etc.) and whether a URL has a bad reputation (e.g., malicious), a good reputation, or an unknown reputation. A posting containing a URL with a bad reputation may be deemed to be a prohibited posting. In that case, the URL may be used as an indicator of a prohibited posting, i.e., any posting with the URL may be deemed to be prohibited. Other features of identified prohibited posting, such as its username (e.g., hashtag), may also be used as indicators of prohibited postings. The category of a feature of a posting (e.g., embedded URLs, linked sites, etc.) may be assigned to the posting.
In one embodiment, the security system 220 includes a machine learning model 246 for correlation, categorization, and identifying indicators of prohibited postings. For example, the content of a posting may be run through the machine learning model 246 to determine the category of the posting, etc. A posting may be deemed prohibited if it belongs to a prohibited category as indicated by one or more computer security policies. The machine learning model 246 may be trained using rules and samples obtained from analyzed posting streams and other sources, such as honeypots. The machine learning model 246 may be created using suitable machine learning algorithm, such as support vector machine (SVM), for example.
Sites (i.e., websites and other computer nodes on the Internet) that are linked in a posting may be followed to determine a category of the posting. For example, a posting with a link to a political site may be categorized as “controversial” (and prohibited in an enterprise environment) because political sites typically have different views on politics. A posting may be in one or more categories and may be deemed to be prohibited if any of the categories is prohibited by a computer security policy.
The security system 220 may set the prohibited postings based on its processing of the dynamic postings (arrow 204). In one embodiment, the security system 220 informs the content enforcement endpoint 241 the particular postings that are prohibited and/or the indicators of prohibited postings. For example, the content enforcement endpoint 241 may receive URLs and/or usernames of prohibited postings from the security system 220, and block postings that include those URLs and/or usernames. The content enforcement endpoint 241 is thus able to identify and block prohibited postings that are incoming with the posting stream. An IT administrator may also set prohibited postings by indicating them in computer security policies (arrow 205). The content enforcement endpoint 241 allows permitted postings, i.e., postings that are not prohibited, to enter the enterprise network 240 (arrow 206). The process of receiving and analyzing dynamic postings to identify prohibited postings may be performed periodically, e.g., at least every hour or more frequently, to keep up with dynamically changing postings.
In one embodiment, information obtained by analyzing postings of one social network may be employed to control usage of another social network. For example, indicators of prohibited postings identified by processing the dynamic postings of the social network 230-1 may also be employed by the content enforcement endpoint 241 to identify prohibited postings in the posting stream of the social network 230-2 (arrow 207). For example, URLs that indicate prohibited postings on the social network 230-1 may also be used as indicators of prohibited postings on the social network 230-2.
As can be appreciated, the system 200 may be employed to control usage of different social networks. For example, the system 200 may be employed to control usage of the TWITTER social network in the computer network 240 as now explained beginning with
In the example of
The filtered tweets may be received in the security system 200 and indexed into a full text search engine (step 303). For example, the filtered tweets may be indexed in the Elasticsearch search engine or other suitable search engines. The indexed tweets may be queried using the search engine to determine the categories of the tweets (step 304). The categories may include “porn”, “riots”, “violence”, etc. or other categories that are dictated by one or more computer security policies enforced by the content enforcement endpoint 241.
The category of a tweet, and hence its hashtag, may be based on the category of a URL included in the tweet or the category of the TWITTER URL of the end-user who posted the tweet. For example, a tweet may include a shortened URL. To determine the category of the tweet, the shortened URL may be followed to determine the full URL that corresponds to the shortened URL. The category of the full URL may be obtained by consulting a web reputation system, using a machine learning model, by inspection, etc. The distribution channels of the full URL may also be investigated to determine the category of the full URL. The TWITTER handles (i.e., usernames) of tweets that contain the URLs may also be investigated to determine the category of a tweet. For example, some famous sports personalities have verified TWITTER handles, making tweets by those TWITTER handles likely to be in the “sports” category. The same procedure may be performed for other verified TWITTER handles.
Information obtained by categorizing the indexed tweets may be used to identify prohibited tweets. In the example of
The system 200 may also be adapted to control usage of the REDDIT social network in the computer network 240. The REDDIT social network, having a primary website address of <<www.reddit.com>>, also has a lot of dynamic content in, for example, live threads and sub-groups that are called subreddits. Live threads primarily capture live events as they are happening, similar to the “trending” topics on the TWITTER and FACEBOOK social networks. Live threads typically go to background (while still staying alive) as more and more new threads arise. The contents of a live thread may be accessed from <<http://www.reddit.comni<livethread-name>>>.
Subreddits, like live threads, are also dynamic and created by end-users. A subreddit gains prominence as more and more end-users post to it. The contents of a subreddit may be accessed from <<http://wwvv.reddit.com/d<subreddit-name>>>. The number of subreddits is relatively large and the categorizations of sub-reddits are very broad. For example,
The categorization problem involving the REDDIT social network becomes worse when contents of so-called “big events” show up on the REDDIT social network. One example is the case of stolen private photos from an iCloud hack. The stolen private photos were first posted on the 4chan site but were viewed the most on the REDDIT social network. The corresponding dynamically created subreddit was <<http://www.reddit.com/r/thefappening>>, which was accessible for a week before being disabled. Because the subreddit was not properly categorized, some end-users on enterprise networks were able to view the subreddit even with computer security policies being enforced.
In the example of
Methods and systems for controlling social network usage in enterprise environments have been disclosed. While specific embodiments of the present invention have been provided, it is to be understood that these embodiments are for illustration purposes and not limiting. Many additional embodiments will be apparent to persons of ordinary skill in the art reading this disclosure.
Number | Name | Date | Kind |
---|---|---|---|
7021534 | Kiliccote | Apr 2006 | B1 |
7590707 | McCloy, III et al. | Sep 2009 | B2 |
7802298 | Hong et al. | Sep 2010 | B1 |
7854001 | Chen et al. | Dec 2010 | B1 |
7984500 | Khanna et al. | Jul 2011 | B1 |
8381292 | Warner et al. | Feb 2013 | B1 |
8468597 | Warner et al. | Jun 2013 | B1 |
8495735 | Warner et al. | Jul 2013 | B1 |
9183259 | Marra | Nov 2015 | B1 |
20050160330 | Embree et al. | Jul 2005 | A1 |
20060041508 | Pham et al. | Feb 2006 | A1 |
20060064374 | Helsper et al. | Mar 2006 | A1 |
20060070126 | Grynberg et al. | Mar 2006 | A1 |
20060080735 | Brinson et al. | Apr 2006 | A1 |
20060101120 | Helsper et al. | May 2006 | A1 |
20060123464 | Goodman et al. | Jun 2006 | A1 |
20060123478 | Rehfuss et al. | Jun 2006 | A1 |
20060168066 | Helsper et al. | Jul 2006 | A1 |
20070112814 | Chesshire | May 2007 | A1 |
20070118904 | Goodman et al. | May 2007 | A1 |
20070282739 | Thomsen | Dec 2007 | A1 |
20080028444 | Loesch et al. | Jan 2008 | A1 |
20080034073 | McCloy et al. | Feb 2008 | A1 |
20080082662 | Dandliker et al. | Apr 2008 | A1 |
20080133540 | Hubbard et al. | Jun 2008 | A1 |
20100095378 | Oliver et al. | Apr 2010 | A1 |
20110167474 | Sinha et al. | Jul 2011 | A1 |
20120222111 | Oliver et al. | Aug 2012 | A1 |
20120227104 | Sinha et al. | Sep 2012 | A1 |
20120239763 | Musil | Sep 2012 | A1 |
20130138735 | Kanter | May 2013 | A1 |
20130159417 | Meckler | Jun 2013 | A1 |
20130298038 | Spivack | Nov 2013 | A1 |
20150363796 | Lehman | Dec 2015 | A1 |
20160142358 | Zunger | May 2016 | A1 |
20160171109 | Gnanasekaran | Jun 2016 | A1 |
20160188597 | Moore | Jun 2016 | A1 |
20160277349 | Bhatt | Sep 2016 | A1 |
20170046401 | Schenk | Feb 2017 | A1 |
20170206210 | Goikhman | Jul 2017 | A1 |
Entry |
---|
Seems Somebody is Clicking on That Spam—New York Times, Jul. 3, 2006, 1 sheet [retrieved on Jun. 2, 2014], retrieved from the internet: http://www.nytimes.com/2006/07/03/technology/03drill.html. |
Does the Twitter Follower Scam Actually Work / Security Intelligence Blog / Trend Micro, Jan. 30, 2014, 4 sheets [retrieved on Jun. 2, 2014], retrieved from the internet: http://blog.trendmicro.com/trendlabs-security-intelligence/does-tie-twitter-follower-scam-actually-work/. |
Clique problem—Wikipedia, the free encyclopedia, 11 sheets [retrieved on Jun. 2, 2014], retrieved from the internet: http://en.wikipedia.org/wiki/Clique_problem. |
Bipartite graph—Wikipedia, the free encyclopedia, 9 sheets [retrieved on Jun. 2, 2014], retrieved from the internet: http://en.wikipedia.org/wiki/Bipartite_graph. |
Spam ROI: Profit on 1 in 12.5m Response Rate, Nov. 11, 2008, 5 sheets [retrieved on Jun. 2, 2014], retrieved from the Internet: http://www.sitepoint.com/spam-roi-profit-on-1-in-125m-response-rate/. |
Trend Micro Simply Security Checking Identities in Social Network Friend Requests, Feb. 28, 2013, 5 sheets [retrieved on Jun. 2, 2014], retrieved from the internet: http://blog.trendmicro.com/checking-identities-in-facebook-friend-request/. |
Facebook to put Report Abuse button at fingertips of bullying victims / Naked Security, Nov. 6, 2013, 3 sheets [retrieved on Jun. 2, 2014], retrieved from the internet: http://nakedsecurity.sophos.com/2013/11/06/facebook-to-put-report-abuse-button-at-fingertips-of-bullying-victims/. |
Twitter rolls out ‘report abuse’ button for individual tweets: will you use it?, 4 sheets [retrieved on Jun. 2, 2014], retrieved from the internet: http://www.theguardian.com/technology/blog/2013/aug/30/twitter-report-abuse-button. |
Honeypot (computing)—Wikipedia, the free encyclopedia, 4 sheets [retrieved on Jun. 2, 2014], retrieved from the Internet: http://en.wikipedia.org/wiki/Honeypot_(computing). |
URL shortening-Abuse—Wikipedia, the free encyclopedia, 1 sheet [retrieved on Jun. 5, 2014], retrieved from the Internet: http://en.wikipedia.org/wiki/URL_shortening_-_Abuse. |
URL shortening—Wikipedia, the free encyclopedia, 7 sheets [retrieved on Jun. 5, 2014], retrieved from the internet: http://en.wikipedia.org/wiki/URL_shortening. |
URL redirection—Wikipedia, the free encyclopedia,11 sheets [retrieved on Jun. 5, 2014], retrieved from the internet: http://en.wikipedia.org/wiki/URL_redirection. |
E. Zangerle and G. Specht “Sorry, I was hacked”: A Classification of Compromised Twitter Accounts, Mar. 2014, 7 sheets, retrieved from the internet: http://www.evazangerle.at/wp-content/papercite-data/pdf/sac14.pdf. |
J. Xiang, C. Guo and A. Aboulnaga “Scalable Maximum Clique Computation Using MapRedue”, Jan. 2014, 12 sheets, retrieved from the internet: https://cs.uwaterloo.ca/˜ashraf/pubs/icde13maxclique.pdf. |
C. Grier, K. Thomas, V. Paxson and M. Zhang “@spam: The Underground on 140 Characters or Less”, Oct. 2010, 11 sheets, In Proceeding of the 17th ACM Conference on Computer and Communications Security, retrieved from the internet: http://www.icir.org/vern/papers/ccs2010-twitter-spam.pdf. |
C. Kanich, C. Kreibich, K. Levchenko, B. Enright, G. Voelker, V. Paxson and S. Savage “Spamalytics: An Emperical Analysis of Spam Marketing Conversion” 2008, 12 sheets, In Proceeding of the 15th ACM Conference on Computer and Communications Security, retrieved from the internet: http://www.icsi.berkeley.edu/pubs/networking/2008-ccs-spamalytics.pdf. |
M. Egele, G. Stringhini, C. Kruegel and G. Vigna “COMPA: Detecting Comprimised Accounts on Social Networks”, 2013, 17 sheets, retrieved from the internet: http://www.cs.ucsb.edu/˜gianluca/papers/thjp-ndss13.pdf. |
Michael Steven Svendsen “Mining maximal cliques from large graphs using MapReduce”, 2012, 45 sheets, retrieved from the internet: http://lib.dr.iastate.edu/cgi/viewcontent.cgi?article=3631&context=etd. |
Yun-Chian Cheng “Hadoop Sucess Stories in Trend Micro SPN”, Oct. 2012, 30 sheets, retrieved from the internet: http://www.gwms.com.tw/TREND_HadoopinTaiwan2012/1002download/04.pdf. |
Hung-Tsai Su, S. Tsao, W. Chu and R. Liao “Mining Web Browsing Log by Using Relaxed Biclique Enumeration Algorithm in MapReduce”, 2012, 47 sheets, vol. 3, IEEE/WIC/ACM International Conferences on Web Intelligence and Intelligent Agent Technology. |
Twitter Help Center—The Twitter Rules, 4 sheets (retrieved on Oct. 17, 2014), retrieved from the internet: https://support.twitter.com/articles/18311-the-twitter-rules. |