WO2005027016A2 - Fraudulent message detection - Google Patents

Fraudulent message detection Download PDF

Info

Publication number
WO2005027016A2
WO2005027016A2 PCT/US2004/025438 US2004025438W WO2005027016A2 WO 2005027016 A2 WO2005027016 A2 WO 2005027016A2 US 2004025438 W US2004025438 W US 2004025438W WO 2005027016 A2 WO2005027016 A2 WO 2005027016A2
Authority
WO
WIPO (PCT)
Prior art keywords
message
classifying
recited
identifying
reference points
Prior art date
Application number
PCT/US2004/025438
Other languages
French (fr)
Other versions
WO2005027016A3 (en
Inventor
Jonathan J. Oliver
Scott D. Eikenberry
Gleb Budman
Bryan Kim
Original Assignee
Mailfrontier, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mailfrontier, Inc. filed Critical Mailfrontier, Inc.
Publication of WO2005027016A2 publication Critical patent/WO2005027016A2/en
Publication of WO2005027016A3 publication Critical patent/WO2005027016A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/12Applying verification of the received information
    • H04L63/123Applying verification of the received information received data contents, e.g. message integrity
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/50Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
    • G06F21/51Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems at application loading time, e.g. accepting, rejecting, starting or inhibiting executable software based on integrity or source reliability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/14Network architectures or network communication protocols for network security for detecting or protecting against malicious traffic
    • H04L63/1441Countermeasures against malicious traffic

Definitions

  • the present invention relates generally to electronic messages. More specifically, classifying messages is disclosed.
  • spam filtering software used to detect junk electronic messages
  • spam filtering software detects that a message is junk and places it into a spam folder, because the message appears to be authentic, the recipient may reclassify the message and respond to it. It would be desirable to have a technique that would identify the fraudulent messages and prevent users from becoming victims of such scaixis.
  • Figure 1 is a system diagram illustrating a system used to detect fraud messages according to one embodiment.
  • Figure 2 is a flowchart illustrating a message classification process according to one embodiment.
  • Figure 3 is a flowchart illustrating a message classification process according to another embodiment.
  • the invention can be implemented in numerous ways, including as a process, an apparatus, a system, a composition of matter, a computer readable medium such as a computer readable storage medium or a computer network wherein program instructions are sent over optical or electronic communication links.
  • these implementations, or any other form that the invention may take, may be referred to as techniques.
  • the order of the steps of disclosed processes may be altered within the scope of the invention.
  • the technique uses reference points in the message to classify the message.
  • the message is determined to be a phish message if divergent reference points are detected, h some embodiments, fraud indicators are identified in the message and a statistical analysis is used to determine whether the message is fraudulent.
  • FIG. 1 is a system diagram illustrating a system used to detect fraud messages according to one embodiment.
  • Incoming messages such as message 100 are examined by fraud detection engine 102.
  • the fraud detection engine may be a standalone software and/or hardware component, or integrated into a spam filtering system, mail system, or any other appropriate software code and/or hardware devices. It may be deployed on client devices such as personal computers as well as message gateways used to filter messages at an enterprise level.
  • the fraud detection engine analyzes the message and determines whether it is a fraudulent message. Details of the detection process are discussed below.
  • a thumbprint or signature of the message is optionally computed.
  • the thumbprint may be computed using techniques such as applying a hash function to the content of the message.
  • the thumbprint may be stored in a database 104 to more efficiently identify this fraudulent message in the future.
  • the thumbprint database may be shared among users on a network.
  • the body of tJhis type of messages usually includes reference points to a legitimate source, such as URL links to logo images or web pages of well-known company, e-mail addresses with well- known domain name, as well as any other contact or identity information associated with well-known entities.
  • a reference point refers to information in the message that can be used to contact the referenced entity or establish the identity of the referenced entity;
  • a legitimate source refers to an entity that can " be verified as being reputable.
  • Some of the fraudulent messages include a reference point to a questionable source or entity whose reputation cannot be easily verified.
  • phish messages usually include in the message body at least one reference point to a source that is legitimate and at least one reference point to a questionable source.
  • the characteristic of having a legitimate reference point and a questionable reference point is referred to as divergence, and such reference points are referred to as divergent reference points. Divergence can be used in some embodiments to identify phish messages.
  • FIG. 2 is a flowchart illustrating a message classification process according to one embodiment, hi this example, reference points are extracted from a message (200).
  • the reference point to the questionable source is disguised to trick the reader.
  • the URL http://scgi.ebay.com@%32%31%31%2E%33%37%2E%31%38%35%2E%33%30:% 37%33%30%31 appears to be a link to "scgi.ebay.com” to the casual reader, but it actually links to a web page at a server with the IP address of 211.37.185.30, on port 7301.
  • both the reference point to scgi.ebay.com and the reference point to the numbered server are extracted.
  • the reference points are then classified (202).
  • the reference points are looked up in a database that includes references of legitimate sources. There are various ways to establish and maintain the database.
  • users on a network can collaborate and vote on various reference points to determine whether they are legitimate, and the results of the collaborative effort are stored in the database; in some embodiments, collections of spam messages and good messages can be analyzed and the reference points in the messages can be extracted and stored to the database according to the message classification; in some embodiments, a registration process can be used to collect information from legitimate businesses and organizations; in some embodiments, existing databases of good reference points such as classified URLs, domains and phone number can be imported.
  • a reference point is classified as legitimate if the reference point or the source referenced can be found in the database; otherwise, the reference point is classified as unknown, hi some embodiments, the reference point may also be looked up in a database of known offenders. A reference point found in this database may be classified as bad. Both unknown and bad reference points are classified as questionable reference points in some embodiments.
  • the message is then examined for divergent reference points (204). In some embodiments, different types of reference points are counted, and if there is at least one good reference and at least one questionable reference in the message body, then the message is deemed to have divergent reference points. It is then determined whether the message has divergent reference points (206). A message that does not have divergent reference point is less lilcely to be a phish message and is processed as usual (208). In some embodiments, the message is delivered to the recipient. In some embodiments, the message is further processed by spam classifiers. If, however, the message is determined to have divergent reference points, it is potentially a phish message and is handled appropriately (210). The handling of the message may vary for different implementations.
  • the message may be deleted or quarantined.
  • the recipient may be alerted about the phish message.
  • the alert can take many forms; for example, an indicator such as "[fraud]" may be added to the subject line, the recipient may be notified that a message has been quarantined or deleted, etc.
  • the recipient may be given detailed explanation of why the message is likely to be fraudulent. For example, both the legitimate source from which the message appears to be originated and the questionable source designed to misguide the recipient may be clarified for the user.
  • a statistical approach is used to classify phish messages as well as other types of fraud messages.
  • Figure 3 is a flowchart illustrating a message classification process according to another embodiment.
  • fraud indicators are used to classify fraudulent messages.
  • Fraud indicators are patterns that are characteristic to fraudulent messages. For example, having divergent reference points is one type of fraud indicator. More details of fraud indicators are discussed below.
  • the fraud indicator is identified by parsing the message and locating certain patterns (300). A statistical analysis is then applied based on the fraud indicators (302). There are various statistical techniques applicable for the purposes of identifying fraud messages.
  • the number of fraud indicators is counted and the likelihood of the message being a fraudulent message is computed based on the number of fraud indicators.
  • Other statistical approaches for analyzing the indicators include neural networks, decision tree and classification tree techniques, logistic regression, as well as any other techniques or combination of techniques for analyzing features of the message according to a statistical model.
  • the message is determined to be fraudulent if the number of fraudulent indicators exceeds a threshold.
  • the statistical analysis computes a score that is used to determine whether the message is fraudulent.
  • the statistical analysis assigns weights to different indicators and computes a weighted score used to determine whether the message is fraudulent.
  • the message is delivered to the recipient, h some embodiments, the message is further tested using other spam filtering techniques. If the message is determined to be a fraudulent message, it is handled appropriately so that the recipient is prevented from responding to the fraudulent message (308).
  • the message may be deleted or quarantined in some embodiments; alerts and explanations may be sent to the recipient in some embodiments.
  • a probability of being fraudulent is assigned to the message and the message is processed according to the probability.
  • the statistical message classification process shown in Figure 3 may be used in conjunction with the message classification process shown in Figure 2.
  • a statistical technique is applied to further ascertain the nature of the message.
  • fraud indicators including raw Internet protocol (IP) addresses, non-standard encoding of URL's or addresses, links with user names embedded, misleading links, mismatched link names, message including forms, forms with certain special requests, suspect content, etc.
  • raw IP addresses may be used as a type of fraud indicator in some embodiments.
  • the identification of a raw IP address may be achieved by, for example, locating a string with a pattern of four numbers separated by periods.
  • non-standard encoding is used as a type of fraud indicator.
  • Fraudulent messages sometimes use non-standard encoding of URL to obfuscate their addresses, such as the URL http://scgi.ebay.com@%32%31%31%2E%33%37%2E%31%38%35 %2E%33%30:% 37%33%30%31 shown in a previous example.
  • There are a variety of ways for detennining whether non-standard encoding is use in a message including finding % signs in URL links, using raw IP numbers in decimal, dword, octal or hexadecimal format, etc.
  • URL links embedded with user names are used as a type of fraud indicator since fraudulent messages sometimes use this type of links.
  • http://USER@netgo.port5.com has a user name embedded in it. This type of fraud indicator may be identified by searching for an "@" symbol after "http://" or "https://”.
  • URL links embedded with misleading links are used as a type of fraud indicator since fraudulent messages sometimes include this type of URL.
  • a misleading link disguises the questionable source by incorporating a reference to a well-known source. For example, http://www.paypal.com@netgo.port5. com/index. php includes a user name "www.paypal.com" that has the format of a domain name.
  • This type of fraud indicator may be identified in some embodiments by searching for an "@” symbol followed by a domain name, where the text prior to the "@" symbol includes a substring such as ".com", “.net”, "www", etc.
  • mismatched link names are used as a type of fraud indicator. For example, a hyper text markup language (HTML) document with the URL
  • Form requesting special information such as personal information is another type of fraud indicator that is used in some embodiments.
  • the indicator is present if the message is determined to include a form, and the content of the form starting from the beginning " ⁇ FORM>" tag to the end " ⁇ /FORM>” tag (or the end of the message if the end " ⁇ /FORM>” tag is not present) includes requests for personal information.
  • Whether the form is requesting personal information may be determined by looking up key terms such as "Password”, “Credit card”, “Social security number”, “Expiration date”, “PIN”, and any other terms pertaining to personal information.
  • suspect content is used as a type of fraud indicator.
  • Suspect content may be detected by searching for certain key words or phrases such as "verification of your account”, “click on the link below to login”, “fraud alert”, “please confirm your account ownership by entering the information below”, etc.
  • a technique for classifying a message has been disclosed.
  • the technique is applicable to phish messages as well as other types of fraud messages.
  • the technique may also be applicable to messages used to commit brand fraud, where the recipient is shown a product of a certain brand, but directed to a website unrelated to the brand-name product to purchase a different product.

Abstract

A technique for classifying a message (100) is disclosed. In some embodiments, the technique comprises extracting a plurality of reference points, classifying the plurality of reference points, and detecting that the message (100) is a phish message based on the classified reference points. In some embodiments, the technique comprises identifying a plurality of fraud indicators (104) in the message (100), applying a statistical analysis (102) on the plurality of fraud indicators (104); and determining whether the message (100) is a fraudulent message based on the analysis (102).

Description

FRAUDULENT MESSAGE DETECTION
CROSS REFERENCE TO RELATED APPLICATIONS
This application claims priority to U.S. Provisional Patent Application No. 60/501,481 (Attorney Docket No. MAILP010+) entitled FRAUDULENT MESSAGE DETECTION filed September 8, 2003, which is incorporated herein by reference for all purposes.
FIELD OF THE INVENTION
The present invention relates generally to electronic messages. More specifically, classifying messages is disclosed.
BACKGROUND OF THE INVENTION
There has been an increasing number of Internet scams designed to trick people into revealing personal information such as credit card numbers, passwords, social security numbers, etc. to unscrupulous individuals who intend to use the information for fraudulent purposes. These scams, also known as "phishing", usually involve sending electronic messages such as e-mail to unsuspecting users. These messages are designed to appear as though they come from a reputable source, for example e-commerce companies, banks, credit card companies or other well-known people or institutions. The messages are often disguised as security alerts, account maintenance or renewal notices. The messages typically instruct the recipients to enter their personal information in the message, or direct the recipients to a website operated by the fraud message sender where the information may be entered. Since the messages typically appear to have a legitimate purpose, they often escape the detection of the conventional filtering software used to detect junk electronic messages (commonly referred to as "spam"). In some cases, even when the spam filtering software detects that a message is junk and places it into a spam folder, because the message appears to be authentic, the recipient may reclassify the message and respond to it. It would be desirable to have a technique that would identify the fraudulent messages and prevent users from becoming victims of such scaixis.
BRIEF DESCRIPTION OF THE DRAWINGS
Various embodiments of the invention are disclosed in the following detailed description and the accompanying drawings.
Figure 1 is a system diagram illustrating a system used to detect fraud messages according to one embodiment.
Figure 2 is a flowchart illustrating a message classification process according to one embodiment. Figure 3 is a flowchart illustrating a message classification process according to another embodiment.
DETAILED DESCRIPTION
The invention can be implemented in numerous ways, including as a process, an apparatus, a system, a composition of matter, a computer readable medium such as a computer readable storage medium or a computer network wherein program instructions are sent over optical or electronic communication links. In this specification, these implementations, or any other form that the invention may take, may be referred to as techniques. In general, the order of the steps of disclosed processes may be altered within the scope of the invention.
A detailed description of one or more embodiments of the invention is provided below along with accompanying figures that illustrate the principles of the invention. The invention is described in connection with such embodiments, but the invention is not limited to any embodiment. The scope of the invention is limited only by the claims and the invention encompasses numerous alternatives, modifications and equivalents. Numerous specific details are set forth in the following description in order to provide a thorough understanding of the invention. These details are provided for the purpose of example and invention may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the invention has not been described in detail so that the invention is not unnecessarily obscured. A technique for classifying a message is disclosed. In some embodiments, the technique uses reference points in the message to classify the message. In some embodiments, the message is determined to be a phish message if divergent reference points are detected, h some embodiments, fraud indicators are identified in the message and a statistical analysis is used to determine whether the message is fraudulent.
Figure 1 is a system diagram illustrating a system used to detect fraud messages according to one embodiment. Incoming messages such as message 100 are examined by fraud detection engine 102. The fraud detection engine may be a standalone software and/or hardware component, or integrated into a spam filtering system, mail system, or any other appropriate software code and/or hardware devices. It may be deployed on client devices such as personal computers as well as message gateways used to filter messages at an enterprise level. The fraud detection engine analyzes the message and determines whether it is a fraudulent message. Details of the detection process are discussed below. In some embodiments, once a fraudulent message is detected, a thumbprint or signature of the message is optionally computed. The thumbprint may be computed using techniques such as applying a hash function to the content of the message. The thumbprint may be stored in a database 104 to more efficiently identify this fraudulent message in the future. In some embodiments, the thumbprint database may be shared among users on a network.
To make the fraudulent message appear authentic, the body of tJhis type of messages usually includes reference points to a legitimate source, such as URL links to logo images or web pages of well-known company, e-mail addresses with well- known domain name, as well as any other contact or identity information associated with well-known entities. As used herein, a reference point refers to information in the message that can be used to contact the referenced entity or establish the identity of the referenced entity; a legitimate source refers to an entity that can "be verified as being reputable. In order for the fraudulent message to fulfill its intended purpose, it should provide the recipient with ways to supply the desired information to the sender. Some of the fraudulent messages include a reference point to a questionable source or entity whose reputation cannot be easily verified. The recipient is often instructed to provide his/her personal information to the questionable source by, for example, entering information into the message or by following a link to a website operated by the sender. This type of fraudulent messages, referred to as phish messages, usually include in the message body at least one reference point to a source that is legitimate and at least one reference point to a questionable source. The characteristic of having a legitimate reference point and a questionable reference point is referred to as divergence, and such reference points are referred to as divergent reference points. Divergence can be used in some embodiments to identify phish messages.
Figure 2 is a flowchart illustrating a message classification process according to one embodiment, hi this example, reference points are extracted from a message (200). In some phish messages, the reference point to the questionable source is disguised to trick the reader. For example, the URL http://scgi.ebay.com@%32%31%31%2E%33%37%2E%31%38%35%2E%33%30:% 37%33%30%31 appears to be a link to "scgi.ebay.com" to the casual reader, but it actually links to a web page at a server with the IP address of 211.37.185.30, on port 7301. In some embodiments, both the reference point to scgi.ebay.com and the reference point to the numbered server are extracted. The reference points are then classified (202). In some embodiments, the reference points are looked up in a database that includes references of legitimate sources. There are various ways to establish and maintain the database. In some embodiments, users on a network can collaborate and vote on various reference points to determine whether they are legitimate, and the results of the collaborative effort are stored in the database; in some embodiments, collections of spam messages and good messages can be analyzed and the reference points in the messages can be extracted and stored to the database according to the message classification; in some embodiments, a registration process can be used to collect information from legitimate businesses and organizations; in some embodiments, existing databases of good reference points such as classified URLs, domains and phone number can be imported. In some embodiments, a reference point is classified as legitimate if the reference point or the source referenced can be found in the database; otherwise, the reference point is classified as unknown, hi some embodiments, the reference point may also be looked up in a database of known offenders. A reference point found in this database may be classified as bad. Both unknown and bad reference points are classified as questionable reference points in some embodiments.
In the example shown in Figure 2, the message is then examined for divergent reference points (204). In some embodiments, different types of reference points are counted, and if there is at least one good reference and at least one questionable reference in the message body, then the message is deemed to have divergent reference points. It is then determined whether the message has divergent reference points (206). A message that does not have divergent reference point is less lilcely to be a phish message and is processed as usual (208). In some embodiments, the message is delivered to the recipient. In some embodiments, the message is further processed by spam classifiers. If, however, the message is determined to have divergent reference points, it is potentially a phish message and is handled appropriately (210). The handling of the message may vary for different implementations. In some embodiments, the message may be deleted or quarantined. In some embodiments, the recipient may be alerted about the phish message. The alert can take many forms; for example, an indicator such as "[fraud]" may be added to the subject line, the recipient may be notified that a message has been quarantined or deleted, etc. In some embodiments, the recipient may be given detailed explanation of why the message is likely to be fraudulent. For example, both the legitimate source from which the message appears to be originated and the questionable source designed to misguide the recipient may be clarified for the user. In some embodiments, a statistical approach is used to classify phish messages as well as other types of fraud messages. Figure 3 is a flowchart illustrating a message classification process according to another embodiment. In this example, fraud indicators are used to classify fraudulent messages. Fraud indicators are patterns that are characteristic to fraudulent messages. For example, having divergent reference points is one type of fraud indicator. More details of fraud indicators are discussed below. In some embodiments, the fraud indicator is identified by parsing the message and locating certain patterns (300). A statistical analysis is then applied based on the fraud indicators (302). There are various statistical techniques applicable for the purposes of identifying fraud messages. In some embodiments, the number of fraud indicators is counted and the likelihood of the message being a fraudulent message is computed based on the number of fraud indicators. Other statistical approaches for analyzing the indicators include neural networks, decision tree and classification tree techniques, logistic regression, as well as any other techniques or combination of techniques for analyzing features of the message according to a statistical model.
It is then determined whether the message is a fraudulent message (304). In some embodiments, the message is determined to be fraudulent if the number of fraudulent indicators exceeds a threshold. In some embodiments, the statistical analysis computes a score that is used to determine whether the message is fraudulent. In some embodiments, the statistical analysis assigns weights to different indicators and computes a weighted score used to determine whether the message is fraudulent.
If the statistical analysis detennines that the message is not a fraud lent message, it is processed as usual (306). h some embodiments, the message is delivered to the recipient, h some embodiments, the message is further tested using other spam filtering techniques. If the message is determined to be a fraudulent message, it is handled appropriately so that the recipient is prevented from responding to the fraudulent message (308). The message may be deleted or quarantined in some embodiments; alerts and explanations may be sent to the recipient in some embodiments. In some embodiments, a probability of being fraudulent is assigned to the message and the message is processed according to the probability.
The statistical message classification process shown in Figure 3 may be used in conjunction with the message classification process shown in Figure 2. In some embodiments, once it is determined that the message has divergent reference points (206), a statistical technique is applied to further ascertain the nature of the message. There are several fraud indicators that may be used by the statistical analysis, including raw Internet protocol (IP) addresses, non-standard encoding of URL's or addresses, links with user names embedded, misleading links, mismatched link names, message including forms, forms with certain special requests, suspect content, etc.
Some fraudulent messages disguise the URL to the questionable source by using raw IP addresses such as http:// 211.37.185.30:7301. Therefore, raw IP addresses may be used as a type of fraud indicator in some embodiments. The identification of a raw IP address may be achieved by, for example, locating a string with a pattern of four numbers separated by periods.
In some embodiments, non-standard encoding is used as a type of fraud indicator. Fraudulent messages sometimes use non-standard encoding of URL to obfuscate their addresses, such as the URL http://scgi.ebay.com@%32%31%31%2E%33%37%2E%31%38%35 %2E%33%30:% 37%33%30%31 shown in a previous example. There are a variety of ways for detennining whether non-standard encoding is use in a message, including finding % signs in URL links, using raw IP numbers in decimal, dword, octal or hexadecimal format, etc. In some embodiments, URL links embedded with user names are used as a type of fraud indicator since fraudulent messages sometimes use this type of links. For example, http://USER@netgo.port5.com has a user name embedded in it. This type of fraud indicator may be identified by searching for an "@" symbol after "http://" or "https://". In some embodiments, URL links embedded with misleading links are used as a type of fraud indicator since fraudulent messages sometimes include this type of URL. A misleading link disguises the questionable source by incorporating a reference to a well-known source. For example, http://www.paypal.com@netgo.port5. com/index. php includes a user name "www.paypal.com" that has the format of a domain name. Without closely inspecting the link, the recipient is likely to mistake www.paypal.com as the linked website rather than netgo.port5.com. This type of fraud indicator may be identified in some embodiments by searching for an "@" symbol followed by a domain name, where the text prior to the "@" symbol includes a substring such as ".com", ".net", "www", etc. In some embodiments, mismatched link names are used as a type of fraud indicator. For example, a hyper text markup language (HTML) document with the URL
<A href="http://www. digitalgamma.com/fraud.htrnl"> http://www.BestBuy.com/fraud_department.html <7A> actually links to the domain "www.digitalgamma.com"; however, the standard techniques for rendering HTML usually displays the link to the recipient as if it is a link to "www.BestBuy.com". This type of fraud indicator may be detected in some embodiments by comparing the actual link and the displayed link. Forms in messages are used as a type of fraud indicator in some embodiments.
Although forms are not commonly used in legitimate messages, fraudulent messages sometimes include forms in order to obtain information from the recipient. Thus, in some embodiments, the message is inspected for the <FORM action=...> tag that is embedded in the message to determine whether the fraud indicator is present. Form requesting special information such as personal information is another type of fraud indicator that is used in some embodiments. The indicator is present if the message is determined to include a form, and the content of the form starting from the beginning "<FORM>" tag to the end "</FORM>" tag (or the end of the message if the end "</FORM>" tag is not present) includes requests for personal information. Whether the form is requesting personal information may be determined by looking up key terms such as "Password", "Credit card", "Social security number", "Expiration date", "PIN", and any other terms pertaining to personal information.
In some embodiments, suspect content is used as a type of fraud indicator. Suspect content may be detected by searching for certain key words or phrases such as "verification of your account", "click on the link below to login", "fraud alert", "please confirm your account ownership by entering the information below", etc.
A technique for classifying a message has been disclosed. The technique is applicable to phish messages as well as other types of fraud messages. For example, the technique may also be applicable to messages used to commit brand fraud, where the recipient is shown a product of a certain brand, but directed to a website unrelated to the brand-name product to purchase a different product.
Although the foregoing embodiments have been described in some detail for purposes of clarity of understanding, the invention is not limited to the details provided. There are many alternative ways of implementing the invention. The disclosed embodiments are illustrative and not restrictive.
WHAT IS CLAIMED IS:

Claims

1. A method for classifying a message, comprising: extracting a plurality of reference points; classifying the plurality of reference points; and detecting that the message is a phish message based on the classified reference points.
2. A method for classifying a message as recited in Claim 1, wherein classifying the plurality of reference points including looking up the plurality of reference points in a database.
3. A method for classifying a message as recited in Claim 1, wherein detecting that the message is a phish message includes determining that the message includes divergent reference points.
4. A method for classifying a message as recited in Claim 1, wherein detecting that the message is a phish message includes determining that the plurality of reference points includes a first reference point to a first source and a second reference point to a second source.
5. A method for classifying a message as recited in Claim 1, wherein detecting that the message is a phish message includes determining that the plurality of reference points includes a first reference point to a legitimate source and a second reference point to a questionable source.
6. A method for classifying a message as recited in Claim 1, wherein detecting that the message is a phish message includes determining that the plurality of reference points includes a first reference point to a first source and a second reference point to a second source, and the second reference point is intended to appear as a reference to the first source.
7. A method for classifying a message as recited in Claim 1, further comprising computing a thumbprint of the message and storing the thumbprint to a database.
8. A method for classifying a message as recited in Claim 1, further comprising computing a thumbprint of the message and storing the thumbprint to a database; wherein the database is shared.
9. A method for classifying a message as recited in Claim 1, further comprising identifying a plurality of fraud indicators and applying a statistical analysis on the plurality of fraud indicators.
10. A method for classifying a message as recited in Claim 1, further comprising quarantining the message.
11. A method for classifying a message as recited in Claim 1, further comprising deleting the message.
12. A method for classifying a message as recited in Claim 1, further comprising providing an alert to a recipient of the message.
13. A method for classifying a message as recited in Claim 1, further comprising providing an alert to a recipient indicating that the message is a phish message.
14. A method for classifying a message as recited in Claim 1, further comprising providing an explanation of the phish message to a recipient.
15. A method for classifying a message, comprising: identifying a plurality of fraud indicators in the message; applying a statistical analysis on the plurality of fraud indicators; and determining whether the message is a fraudulent message based on the analysis.
16. A method for classifying a message as recited in Claim 15, wherein identifying the plurality of fraud indicators includes identifying a raw Internet protocol (IP) address.
17. A method for classifying a message as recited in Claim 15, wherein identifying the plurality of fraud indicators includes identifying non-standard encoding in the message.
18. A method for classifying a message as recited in Claim 15, wherein identifying the plurality of fraud indicators includes identifying a link with an embedded user name.
19. A method for classifying a message as recited in Claim 15, wherein identifying the plurality of fraud indicators includes identifying a misleading link.
20. A method for classifying a message as recited in Claim 15, wherein identifying the plurality of fraud indicators includes identifying a mismatched link name.
21. A method for classifying a message as recited in Claim 15 , wherein identifying the plurality of fraud indicators includes identifying a form in the message.
22. A method for classifying a message as recited in Claim 15, wherein identifying the plurality of fraud indicators includes identifying a form in the message that requests special infonnation.
23. A method for classifying a message as recited in Claim 15, wherein identifying the plurality of fraud indicators includes identifying suspect content in the message.
24. A method for classifying a message as recited in Claim 15, wherein applying a statistical analysis on the plurality of fraud indicators includes obtaining a score based on the fraud indicators.
25. A system for classifying a message, comprising: a processor configured to extract a plurality of reference points, classify the plurality of reference points, and detect that the message is a phish message based on the classified reference points; and a memory coupled with the processor, wherein the memory is configured to provide the processor with instructions.
26. A computer program product for classifying a message, the computer program product being embodied in a computer readable medium and comprising computer instructions for: extracting a plurality of reference points; classifying the plurality of reference points; and detecting that the message is a phish message based on the classified reference points.
27. A system for classifying a message, comprising: a processor configured to identify a plurality of fraud indicators in the message, apply a statistical analysis on the plurality of fraud indicators and determine whether the message is a fraudulent message based on the analysis; and a memory coupled with the processor, wherein the memory is configured to provide the processor with instructions.
28. A computer program product for classifying a message, the computer program product being embodied in a computer readable medium and comprising computer instractions for: identifying a plurality of fraud indicators in the message; applying a statistical analysis on the plurality of fraud indicators; and determining whether the message is a fraudulent message based on the analysis.
PCT/US2004/025438 2003-09-08 2004-08-06 Fraudulent message detection WO2005027016A2 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US50148103P 2003-09-08 2003-09-08
US60/501,481 2003-09-08
US10/678,965 2003-10-02
US10/678,965 US7451487B2 (en) 2003-09-08 2003-10-02 Fraudulent message detection

Publications (2)

Publication Number Publication Date
WO2005027016A2 true WO2005027016A2 (en) 2005-03-24
WO2005027016A3 WO2005027016A3 (en) 2006-02-23

Family

ID=34316484

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2004/025438 WO2005027016A2 (en) 2003-09-08 2004-08-06 Fraudulent message detection

Country Status (2)

Country Link
US (5) US7451487B2 (en)
WO (1) WO2005027016A2 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2422224A (en) * 2005-01-14 2006-07-19 Streamshield Networks Ltd An anti-phishing system for enhancing network security
JP2007011967A (en) * 2005-07-04 2007-01-18 Nec Corp Mail filtering system, mail filter, and program
EP1801745A1 (en) * 2005-12-14 2007-06-27 Aladdin Knowledge Systems, Ltd. Method and system for blocking phishing scams
EP2017758A1 (en) * 2007-07-02 2009-01-21 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Computer-assisted system and computer-assisted method for content verification
US7975297B2 (en) 2005-08-16 2011-07-05 Microsoft Corporation Anti-phishing protection
US8028335B2 (en) 2006-06-19 2011-09-27 Microsoft Corporation Protected environments for protecting users against undesirable activities
US8984289B2 (en) 2003-09-08 2015-03-17 Sonicwall, Inc. Classifying a message based on fraud indicators

Families Citing this family (167)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8046832B2 (en) * 2002-06-26 2011-10-25 Microsoft Corporation Spam detector with challenges
US7219148B2 (en) * 2003-03-03 2007-05-15 Microsoft Corporation Feedback loop for spam prevention
US7483947B2 (en) 2003-05-02 2009-01-27 Microsoft Corporation Message rendering for identification of content features
US7272853B2 (en) 2003-06-04 2007-09-18 Microsoft Corporation Origination/destination features and lists for spam prevention
US8145710B2 (en) 2003-06-18 2012-03-27 Symantec Corporation System and method for filtering spam messages utilizing URL filtering module
US7711779B2 (en) 2003-06-20 2010-05-04 Microsoft Corporation Prevention of outgoing spam
US9412123B2 (en) 2003-07-01 2016-08-09 The 41St Parameter, Inc. Keystroke analysis
US8271588B1 (en) * 2003-09-24 2012-09-18 Symantec Corporation System and method for filtering fraudulent email messages
US8990928B1 (en) * 2003-12-11 2015-03-24 Radix Holdings, Llc URL salience
US7590694B2 (en) 2004-01-16 2009-09-15 Gozoom.Com, Inc. System for determining degrees of similarity in email message information
US8423471B1 (en) * 2004-02-04 2013-04-16 Radix Holdings, Llc Protected document elements
US8214438B2 (en) 2004-03-01 2012-07-03 Microsoft Corporation (More) advanced spam detection features
US10999298B2 (en) 2004-03-02 2021-05-04 The 41St Parameter, Inc. Method and system for identifying users and detecting fraud by use of the internet
US8918466B2 (en) 2004-03-09 2014-12-23 Tonny Yu System for email processing and analysis
US7644127B2 (en) 2004-03-09 2010-01-05 Gozoom.Com, Inc. Email analysis using fuzzy matching of text
US7631044B2 (en) 2004-03-09 2009-12-08 Gozoom.Com, Inc. Suppression of undesirable network messages
US20060020812A1 (en) * 2004-04-27 2006-01-26 Shira Steinberg System and method of using human friendly representations of mathematical function results and transaction analysis to prevent fraud
US20060090073A1 (en) * 2004-04-27 2006-04-27 Shira Steinberg System and method of using human friendly representations of mathematical values and activity analysis to confirm authenticity
US9384348B2 (en) 2004-04-29 2016-07-05 James A. Roskind Identity theft countermeasures
US7971246B1 (en) * 2004-04-29 2011-06-28 James A. Roskind Identity theft countermeasures
US20070107053A1 (en) * 2004-05-02 2007-05-10 Markmonitor, Inc. Enhanced responses to online fraud
US7870608B2 (en) * 2004-05-02 2011-01-11 Markmonitor, Inc. Early detection and monitoring of online fraud
US8041769B2 (en) * 2004-05-02 2011-10-18 Markmonitor Inc. Generating phish messages
US8769671B2 (en) * 2004-05-02 2014-07-01 Markmonitor Inc. Online fraud solution
US7913302B2 (en) * 2004-05-02 2011-03-22 Markmonitor, Inc. Advanced responses to online fraud
US20070299915A1 (en) * 2004-05-02 2007-12-27 Markmonitor, Inc. Customer-based detection of online fraud
US7992204B2 (en) * 2004-05-02 2011-08-02 Markmonitor, Inc. Enhanced responses to online fraud
US9203648B2 (en) 2004-05-02 2015-12-01 Thomson Reuters Global Resources Online fraud solution
US7457823B2 (en) * 2004-05-02 2008-11-25 Markmonitor Inc. Methods and systems for analyzing data related to possible online fraud
US7941490B1 (en) * 2004-05-11 2011-05-10 Symantec Corporation Method and apparatus for detecting spam in email messages and email attachments
US20050289148A1 (en) * 2004-06-10 2005-12-29 Steven Dorner Method and apparatus for detecting suspicious, deceptive, and dangerous links in electronic messages
US7565445B2 (en) * 2004-06-18 2009-07-21 Fortinet, Inc. Systems and methods for categorizing network traffic content
US7664819B2 (en) * 2004-06-29 2010-02-16 Microsoft Corporation Incremental anti-spam lookup and update service
US8412837B1 (en) 2004-07-08 2013-04-02 James A. Roskind Data privacy
US7904517B2 (en) 2004-08-09 2011-03-08 Microsoft Corporation Challenge response systems
US7660865B2 (en) * 2004-08-12 2010-02-09 Microsoft Corporation Spam filtering with probabilistic secure hashes
US8914309B2 (en) 2004-08-20 2014-12-16 Ebay Inc. Method and system for tracking fraudulent activity
GB2418330B (en) * 2004-09-17 2006-11-08 Jeroen Oostendorp Platform for intelligent Email distribution
US20060070126A1 (en) * 2004-09-26 2006-03-30 Amiram Grynberg A system and methods for blocking submission of online forms.
US8056128B1 (en) * 2004-09-30 2011-11-08 Google Inc. Systems and methods for detecting potential communications fraud
US7630987B1 (en) * 2004-11-24 2009-12-08 Bank Of America Corporation System and method for detecting phishers by analyzing website referrals
US8291065B2 (en) * 2004-12-02 2012-10-16 Microsoft Corporation Phishing detection, prevention, and notification
US20060123478A1 (en) * 2004-12-02 2006-06-08 Microsoft Corporation Phishing detection, prevention, and notification
US7634810B2 (en) * 2004-12-02 2009-12-15 Microsoft Corporation Phishing detection, prevention, and notification
JP2006221242A (en) * 2005-02-08 2006-08-24 Fujitsu Ltd Authentication information fraud prevention system, program, and method
US7590698B1 (en) * 2005-03-14 2009-09-15 Symantec Corporation Thwarting phishing attacks by using pre-established policy files
US20060224677A1 (en) * 2005-04-01 2006-10-05 Baytsp Method and apparatus for detecting email fraud
US7841003B1 (en) * 2005-05-04 2010-11-23 Capital One Financial Corporation Phishing solution method
US8010609B2 (en) * 2005-06-20 2011-08-30 Symantec Corporation Method and apparatus for maintaining reputation lists of IP addresses to detect email spam
US7739337B1 (en) 2005-06-20 2010-06-15 Symantec Corporation Method and apparatus for grouping spam email messages
CA2613083A1 (en) * 2005-07-01 2007-01-11 Markmonitor Inc. Enhanced fraud monitoring systems
US7930353B2 (en) * 2005-07-29 2011-04-19 Microsoft Corporation Trees of classifiers for detecting email spam
US8726369B1 (en) 2005-08-11 2014-05-13 Aaron T. Emigh Trusted path, authentication and data security
US8166547B2 (en) 2005-09-06 2012-04-24 Fortinet, Inc. Method, apparatus, signals, and medium for managing a transfer of data in a data network
US8065370B2 (en) 2005-11-03 2011-11-22 Microsoft Corporation Proofs to filter spam
US20070136573A1 (en) * 2005-12-05 2007-06-14 Joseph Steinberg System and method of using two or more multi-factor authentication mechanisms to authenticate online parties
WO2007070612A2 (en) * 2005-12-14 2007-06-21 Tekelec Methods, systems, and computer program products for detecting and mitigating fraudulent message service message traffic
US11301585B2 (en) 2005-12-16 2022-04-12 The 41St Parameter, Inc. Methods and apparatus for securely displaying digital images
US8938671B2 (en) 2005-12-16 2015-01-20 The 41St Parameter, Inc. Methods and apparatus for securely displaying digital images
US8151327B2 (en) 2006-03-31 2012-04-03 The 41St Parameter, Inc. Systems and methods for detection of session tampering and fraud prevention
US7779465B2 (en) * 2006-05-26 2010-08-17 Microsoft Corporation Distributed peer attack alerting
US20080005249A1 (en) * 2006-07-03 2008-01-03 Hart Matt E Method and apparatus for determining the importance of email messages
US8220047B1 (en) 2006-08-09 2012-07-10 Google Inc. Anti-phishing system and method
US7725585B2 (en) * 2006-08-31 2010-05-25 Red Hat, Inc. Methods and systems for alerting a user interface with full destination information
US20080060062A1 (en) * 2006-08-31 2008-03-06 Robert B Lord Methods and systems for preventing information theft
US8050983B1 (en) 2006-10-31 2011-11-01 Amazon Technologies, Inc. Inhibiting inappropriate communications between users involving tranactions
US9110975B1 (en) 2006-11-02 2015-08-18 Google Inc. Search result inputs using variant generalized queries
US8661029B1 (en) 2006-11-02 2014-02-25 Google Inc. Modifying search result ranking based on implicit user feedback
US8224905B2 (en) 2006-12-06 2012-07-17 Microsoft Corporation Spam filtration utilizing sender activity data
US8938463B1 (en) 2007-03-12 2015-01-20 Google Inc. Modifying search result ranking based on implicit user feedback and a model of presentation bias
US8694374B1 (en) * 2007-03-14 2014-04-08 Google Inc. Detecting click spam
US9092510B1 (en) 2007-04-30 2015-07-28 Google Inc. Modifying search result ranking based on a temporal element of user feedback
US9813431B2 (en) * 2007-05-31 2017-11-07 Red Hat, Inc. Browser initiated reporting of fraud
US8694511B1 (en) 2007-08-20 2014-04-08 Google Inc. Modifying search result ranking based on populations
US20090089859A1 (en) * 2007-09-28 2009-04-02 Cook Debra L Method and apparatus for detecting phishing attempts solicited by electronic mail
US8909655B1 (en) 2007-10-11 2014-12-09 Google Inc. Time based ranking
US7836061B1 (en) * 2007-12-29 2010-11-16 Kaspersky Lab, Zao Method and system for classifying electronic text messages and spam messages
US9015130B1 (en) * 2008-03-25 2015-04-21 Avaya Inc. Automatic adjustment of email filters based on browser history and telecommunication records
US8670334B2 (en) * 2008-06-04 2014-03-11 Cisco Technology, Inc Click quality classification and delivery
US20100042687A1 (en) * 2008-08-12 2010-02-18 Yahoo! Inc. System and method for combating phishing
US8607347B2 (en) * 2008-09-29 2013-12-10 Sophos Limited Network stream scanning facility
US8396865B1 (en) 2008-12-10 2013-03-12 Google Inc. Sharing search engine relevance data between corpora
US9112850B1 (en) 2009-03-25 2015-08-18 The 41St Parameter, Inc. Systems and methods of sharing information through a tag-based consortium
US9009146B1 (en) 2009-04-08 2015-04-14 Google Inc. Ranking search results based on similar queries
US8621614B2 (en) * 2009-05-26 2013-12-31 Microsoft Corporation Managing potentially phishing messages in a non-web mail client context
US8447760B1 (en) 2009-07-20 2013-05-21 Google Inc. Generating a related set of documents for an initial set of documents
US8498974B1 (en) 2009-08-31 2013-07-30 Google Inc. Refining search results
US8566943B2 (en) * 2009-10-01 2013-10-22 Kaspersky Lab, Zao Asynchronous processing of events for malware detection
US8972391B1 (en) 2009-10-02 2015-03-03 Google Inc. Recent interest based relevance scoring
US8874555B1 (en) 2009-11-20 2014-10-28 Google Inc. Modifying scoring data based on historical changes
US8615514B1 (en) 2010-02-03 2013-12-24 Google Inc. Evaluating website properties by partitioning user feedback
US8516100B1 (en) * 2010-02-04 2013-08-20 Symantec Corporation Method and apparatus for detecting system message misrepresentation using a keyword analysis
US8924379B1 (en) 2010-03-05 2014-12-30 Google Inc. Temporal-based score adjustments
US8959093B1 (en) 2010-03-15 2015-02-17 Google Inc. Ranking search results based on anchors
US9623119B1 (en) 2010-06-29 2017-04-18 Google Inc. Accentuating search results
US8832083B1 (en) 2010-07-23 2014-09-09 Google Inc. Combining user feedback
US8862999B2 (en) 2010-11-22 2014-10-14 International Business Machines Corporation Dynamic de-identification of data
US9323948B2 (en) * 2010-12-14 2016-04-26 International Business Machines Corporation De-identification of data
US9002867B1 (en) 2010-12-30 2015-04-07 Google Inc. Modifying ranking data based on document changes
US20120180134A1 (en) * 2011-01-07 2012-07-12 Research In Motion Limited Personal Information Guard
US10754913B2 (en) 2011-11-15 2020-08-25 Tapad, Inc. System and method for analyzing user device information
US8813239B2 (en) 2012-01-17 2014-08-19 Bitdefender IPR Management Ltd. Online fraud detection dynamic scoring aggregation systems and methods
US9473437B1 (en) * 2012-02-13 2016-10-18 ZapFraud, Inc. Tertiary classification of communications
US9633201B1 (en) 2012-03-01 2017-04-25 The 41St Parameter, Inc. Methods and systems for fraud containment
US9521551B2 (en) 2012-03-22 2016-12-13 The 41St Parameter, Inc. Methods and systems for persistent cross-application mobile device identification
US9635067B2 (en) 2012-04-23 2017-04-25 Verint Americas Inc. Tracing and asynchronous communication network and routing method
US20130282844A1 (en) 2012-04-23 2013-10-24 Contact Solutions LLC Apparatus and methods for multi-mode asynchronous communication
WO2014022813A1 (en) 2012-08-02 2014-02-06 The 41St Parameter, Inc. Systems and methods for accessing records via derivative locators
US8566938B1 (en) * 2012-11-05 2013-10-22 Astra Identity, Inc. System and method for electronic message analysis for phishing detection
US9154514B1 (en) 2012-11-05 2015-10-06 Astra Identity, Inc. Systems and methods for electronic message analysis
US9501746B2 (en) * 2012-11-05 2016-11-22 Astra Identity, Inc. Systems and methods for electronic message analysis
WO2014078569A1 (en) 2012-11-14 2014-05-22 The 41St Parameter, Inc. Systems and methods of global identification
US9027128B1 (en) * 2013-02-07 2015-05-05 Trend Micro Incorporated Automatic identification of malicious budget codes and compromised websites that are employed in phishing attacks
US9398038B2 (en) 2013-02-08 2016-07-19 PhishMe, Inc. Collaborative phishing attack detection
US9253207B2 (en) 2013-02-08 2016-02-02 PhishMe, Inc. Collaborative phishing attack detection
US9356948B2 (en) 2013-02-08 2016-05-31 PhishMe, Inc. Collaborative phishing attack detection
US9053326B2 (en) 2013-02-08 2015-06-09 PhishMe, Inc. Simulated phishing attack with sequential messages
US8966637B2 (en) 2013-02-08 2015-02-24 PhishMe, Inc. Performance benchmarking for simulated phishing attacks
US9077744B2 (en) 2013-03-06 2015-07-07 Facebook, Inc. Detection of lockstep behavior
US10902327B1 (en) 2013-08-30 2021-01-26 The 41St Parameter, Inc. System and method for device identification and uniqueness
US10277628B1 (en) 2013-09-16 2019-04-30 ZapFraud, Inc. Detecting phishing attempts
US10694029B1 (en) 2013-11-07 2020-06-23 Rightquestion, Llc Validating automatic number identification data
US9111093B1 (en) 2014-01-19 2015-08-18 Google Inc. Using signals from developer clusters
US9246923B1 (en) 2014-01-19 2016-01-26 Google Inc. Developer risk classifier
US9262629B2 (en) 2014-01-21 2016-02-16 PhishMe, Inc. Methods and systems for preventing malicious use of phishing simulation records
BR112016017972B1 (en) 2014-02-06 2022-08-30 Contact Solutions LLC METHOD FOR MODIFICATION OF COMMUNICATION FLOW
CN105279398A (en) * 2014-05-29 2016-01-27 宇龙计算机通信科技(深圳)有限公司 Method and device for protecting information, and terminal
US10027702B1 (en) 2014-06-13 2018-07-17 Trend Micro Incorporated Identification of malicious shortened uniform resource locators
US10078750B1 (en) 2014-06-13 2018-09-18 Trend Micro Incorporated Methods and systems for finding compromised social networking accounts
US10091312B1 (en) 2014-10-14 2018-10-02 The 41St Parameter, Inc. Data structures for intelligently resolving deterministic and probabilistic device identifiers to device profiles and/or groups
US9166881B1 (en) 2014-12-31 2015-10-20 Contact Solutions LLC Methods and apparatus for adaptive bandwidth-based communication management
US9906539B2 (en) 2015-04-10 2018-02-27 PhishMe, Inc. Suspicious message processing and incident response
US9961090B2 (en) 2015-06-18 2018-05-01 Bank Of America Corporation Message quarantine
WO2017024248A1 (en) 2015-08-06 2017-02-09 Contact Solutions LLC Tracing and asynchronous communication network and routing method
US10057198B1 (en) 2015-11-05 2018-08-21 Trend Micro Incorporated Controlling social network usage in enterprise environments
US9954877B2 (en) 2015-12-21 2018-04-24 Ebay Inc. Automatic detection of hidden link mismatches with spoofed metadata
US10721195B2 (en) 2016-01-26 2020-07-21 ZapFraud, Inc. Detection of business email compromise
US10333950B2 (en) 2016-05-01 2019-06-25 International Business Machines Corporation Defending against malicious electronic messages
US10244109B2 (en) * 2016-07-13 2019-03-26 International Business Machines Corporation Detection of a spear-phishing phone call
US10193923B2 (en) * 2016-07-20 2019-01-29 Duo Security, Inc. Methods for preventing cyber intrusions and phishing activity
US10805314B2 (en) 2017-05-19 2020-10-13 Agari Data, Inc. Using message context to evaluate security of requested data
US11936604B2 (en) 2016-09-26 2024-03-19 Agari Data, Inc. Multi-level security analysis and intermediate delivery of an electronic message
US10880322B1 (en) 2016-09-26 2020-12-29 Agari Data, Inc. Automated tracking of interaction with a resource of a message
US9847973B1 (en) 2016-09-26 2017-12-19 Agari Data, Inc. Mitigating communication risk by detecting similarity to a trusted message contact
US11722513B2 (en) 2016-11-30 2023-08-08 Agari Data, Inc. Using a measure of influence of sender in determining a security risk associated with an electronic message
US11044267B2 (en) 2016-11-30 2021-06-22 Agari Data, Inc. Using a measure of influence of sender in determining a security risk associated with an electronic message
US10715543B2 (en) 2016-11-30 2020-07-14 Agari Data, Inc. Detecting computer security risk based on previously observed communications
US10567430B2 (en) 2016-12-09 2020-02-18 International Business Machines Corporation Protecting against notification based phishing attacks
US11019076B1 (en) 2017-04-26 2021-05-25 Agari Data, Inc. Message security assessment using sender identity profiles
US10699295B1 (en) * 2017-05-05 2020-06-30 Wells Fargo Bank, N.A. Fraudulent content detector using augmented reality platforms
US11102244B1 (en) 2017-06-07 2021-08-24 Agari Data, Inc. Automated intelligence gathering
US11757914B1 (en) 2017-06-07 2023-09-12 Agari Data, Inc. Automated responsive message to determine a security risk of a message sender
US10664656B2 (en) * 2018-06-20 2020-05-26 Vade Secure Inc. Methods, devices and systems for data augmentation to improve fraud detection
US11824870B2 (en) 2018-12-19 2023-11-21 Abnormal Security Corporation Threat detection platforms for detecting, characterizing, and remediating email-based threats in real time
US11050793B2 (en) 2018-12-19 2021-06-29 Abnormal Security Corporation Retrospective learning of communication patterns by machine learning models for discovering abnormal behavior
US11431738B2 (en) 2018-12-19 2022-08-30 Abnormal Security Corporation Multistage analysis of emails to identify security threats
US11601440B2 (en) * 2019-04-30 2023-03-07 William Pearce Method of detecting an email phishing attempt or fraudulent email using sequential email numbering
US11470042B2 (en) 2020-02-21 2022-10-11 Abnormal Security Corporation Discovering email account compromise through assessments of digital activities
RU2758358C2 (en) * 2020-02-26 2021-10-28 Акционерное общество "Лаборатория Касперского" Method for generating signature for spam detection
US11477234B2 (en) 2020-02-28 2022-10-18 Abnormal Security Corporation Federated database for establishing and tracking risk of interactions with third parties
US11252189B2 (en) 2020-03-02 2022-02-15 Abnormal Security Corporation Abuse mailbox for facilitating discovery, investigation, and analysis of email-based threats
WO2021178423A1 (en) 2020-03-02 2021-09-10 Abnormal Security Corporation Multichannel threat detection for protecting against account compromise
WO2021183939A1 (en) 2020-03-12 2021-09-16 Abnormal Security Corporation Improved investigation of threats using queryable records of behavior
EP4139801A1 (en) 2020-04-23 2023-03-01 Abnormal Security Corporation Detection and prevention of external fraud
US11528242B2 (en) 2020-10-23 2022-12-13 Abnormal Security Corporation Discovering graymail through real-time analysis of incoming email
US11687648B2 (en) 2020-12-10 2023-06-27 Abnormal Security Corporation Deriving and surfacing insights regarding security threats
US11831661B2 (en) 2021-06-03 2023-11-28 Abnormal Security Corporation Multi-tiered approach to payload detection for incoming communications

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030088627A1 (en) * 2001-07-26 2003-05-08 Rothwell Anton C. Intelligent SPAM detection system using an updateable neural analysis engine
US20050257261A1 (en) * 2004-05-02 2005-11-17 Emarkmonitor, Inc. Online fraud solution

Family Cites Families (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09311806A (en) 1996-05-24 1997-12-02 Hitachi Ltd Method for detecting illegal update of data
CA2261262C (en) 1996-07-22 2007-08-21 Cyva Research Corporation Personal information security and exchange tool
US5878337A (en) 1996-08-08 1999-03-02 Joao; Raymond Anthony Transaction security apparatus and method
US6167520A (en) * 1996-11-08 2000-12-26 Finjan Software, Inc. System and method for protecting a client during runtime from hostile downloadables
US6154844A (en) * 1996-11-08 2000-11-28 Finjan Software, Ltd. System and method for attaching a downloadable security profile to a downloadable
US6122740A (en) 1996-12-19 2000-09-19 Intel Corporation Method and apparatus for remote network access logging and reporting
US5999807A (en) * 1997-02-28 1999-12-07 Cellular Technical Services Company, Inc. System and method for the verification of authentic telephone numbers in a wireless telephone system
US6336109B2 (en) * 1997-04-15 2002-01-01 Cerebrus Solutions Limited Method and apparatus for inducing rules from data classifiers
US6199102B1 (en) 1997-08-26 2001-03-06 Christopher Alan Cobb Method and system for filtering electronic messages
US6334121B1 (en) 1998-05-04 2001-12-25 Virginia Commonwealth University Usage pattern based user authenticator
US6161130A (en) * 1998-06-23 2000-12-12 Microsoft Corporation Technique which utilizes a probabilistic classifier to detect "junk" e-mail by automatically updating a training and re-training the classifier based on the updated training set
US6112227A (en) 1998-08-06 2000-08-29 Heiner; Jeffrey Nelson Filter-in method for reducing junk e-mail
US6321338B1 (en) 1998-11-09 2001-11-20 Sri International Network surveillance
US6763462B1 (en) * 1999-10-05 2004-07-13 Micron Technology, Inc. E-mail virus detection utility
US6851057B1 (en) * 1999-11-30 2005-02-01 Symantec Corporation Data driven detection of viruses
US6650890B1 (en) 2000-09-29 2003-11-18 Postini, Inc. Value-added electronic messaging services and transparent implementation thereof using intermediate server
US6757830B1 (en) * 2000-10-03 2004-06-29 Networks Associates Technology, Inc. Detecting unwanted properties in received email messages
US6802012B1 (en) * 2000-10-03 2004-10-05 Networks Associates Technology, Inc. Scanning computer files for unwanted properties
CA2354372A1 (en) * 2001-02-23 2002-08-23 Efunds Corporation Electronic payment and authentication system with debit and identification data verification and electronic check capabilities
US7234167B2 (en) * 2001-09-06 2007-06-19 Mcafee, Inc. Automatic builder of detection and cleaning routines for computer viruses
US7096500B2 (en) * 2001-12-21 2006-08-22 Mcafee, Inc. Predictive malware scanning of internet data
US7114185B2 (en) * 2001-12-26 2006-09-26 Mcafee, Inc. Identifying malware containing computer files using embedded text
KR100871581B1 (en) 2002-02-19 2008-12-02 포스티니 코포레이션 E-mail management services
US7458098B2 (en) 2002-03-08 2008-11-25 Secure Computing Corporation Systems and methods for enhancing electronic communication security
US8578480B2 (en) * 2002-03-08 2013-11-05 Mcafee, Inc. Systems and methods for identifying potentially malicious messages
US7516182B2 (en) 2002-06-18 2009-04-07 Aol Llc Practical techniques for reducing unsolicited electronic messages by identifying sender's addresses
US7310660B1 (en) * 2002-06-25 2007-12-18 Engate Technology Corporation Method for removing unsolicited e-mail messages
US20040024639A1 (en) 2002-08-05 2004-02-05 Goldman Phillip Y. Direct marketing management on behalf of subscribers and marketers
US7219239B1 (en) * 2002-12-02 2007-05-15 Arcsight, Inc. Method for batching events for transmission by software agent
US6938167B2 (en) 2002-12-18 2005-08-30 America Online, Inc. Using trusted communication channel to combat user name/password theft
US7069487B2 (en) 2003-02-10 2006-06-27 International Business Machines Corporation Pattern matching to validate and classify computer variables to one or more name ranges
US7406502B1 (en) * 2003-02-20 2008-07-29 Sonicwall, Inc. Method and system for classifying a message based on canonical equivalent of acceptable items included in the message
US7590695B2 (en) 2003-05-09 2009-09-15 Aol Llc Managing electronic messages
US20050108340A1 (en) * 2003-05-15 2005-05-19 Matt Gleeson Method and apparatus for filtering email spam based on similarity measures
US7293063B1 (en) * 2003-06-04 2007-11-06 Symantec Corporation System utilizing updated spam signatures for performing secondary signature-based analysis of a held e-mail to improve spam email detection
US20050108562A1 (en) * 2003-06-18 2005-05-19 Khazan Roger I. Technique for detecting executable malicious code using a combination of static and dynamic analyses
US7155484B2 (en) * 2003-06-30 2006-12-26 Bellsouth Intellectual Property Corporation Filtering email messages corresponding to undesirable geographical regions
US7257842B2 (en) * 2003-07-21 2007-08-14 Mcafee, Inc. Pre-approval of computer files during a malware detection
US7835294B2 (en) * 2003-09-03 2010-11-16 Gary Stephen Shuster Message filtering method
US7451487B2 (en) 2003-09-08 2008-11-11 Sonicwall, Inc. Fraudulent message detection
US7870608B2 (en) * 2004-05-02 2011-01-11 Markmonitor, Inc. Early detection and monitoring of online fraud
US20070107053A1 (en) * 2004-05-02 2007-05-10 Markmonitor, Inc. Enhanced responses to online fraud
US7992204B2 (en) * 2004-05-02 2011-08-02 Markmonitor, Inc. Enhanced responses to online fraud
US7913302B2 (en) * 2004-05-02 2011-03-22 Markmonitor, Inc. Advanced responses to online fraud
US9203648B2 (en) * 2004-05-02 2015-12-01 Thomson Reuters Global Resources Online fraud solution
US20070299915A1 (en) * 2004-05-02 2007-12-27 Markmonitor, Inc. Customer-based detection of online fraud
US8041769B2 (en) * 2004-05-02 2011-10-18 Markmonitor Inc. Generating phish messages
US7941490B1 (en) * 2004-05-11 2011-05-10 Symantec Corporation Method and apparatus for detecting spam in email messages and email attachments
US7660865B2 (en) * 2004-08-12 2010-02-09 Microsoft Corporation Spam filtering with probabilistic secure hashes
US20080086532A1 (en) * 2004-10-04 2008-04-10 Brian Cunningham Method for the Verification of Electronic Message Delivery and for the Collection of Data Related to Electronic Messages Sent with False Origination Addresses
US7574409B2 (en) * 2004-11-04 2009-08-11 Vericept Corporation Method, apparatus, and system for clustering and classification
US8291065B2 (en) * 2004-12-02 2012-10-16 Microsoft Corporation Phishing detection, prevention, and notification
US20060123478A1 (en) * 2004-12-02 2006-06-08 Microsoft Corporation Phishing detection, prevention, and notification
US7634810B2 (en) * 2004-12-02 2009-12-15 Microsoft Corporation Phishing detection, prevention, and notification
US8396927B2 (en) * 2004-12-21 2013-03-12 Alcatel Lucent Detection of unwanted messages (spam)
US20060135132A1 (en) * 2004-12-21 2006-06-22 Lucent Technologies, Inc. Storing anti-spam black lists
US20060224677A1 (en) * 2005-04-01 2006-10-05 Baytsp Method and apparatus for detecting email fraud
US8010609B2 (en) * 2005-06-20 2011-08-30 Symantec Corporation Method and apparatus for maintaining reputation lists of IP addresses to detect email spam
US20070136806A1 (en) * 2005-12-14 2007-06-14 Aladdin Knowledge Systems Ltd. Method and system for blocking phishing scams
US20080059590A1 (en) * 2006-09-05 2008-03-06 Ecole Polytechnique Federale De Lausanne (Epfl) Method to filter electronic messages in a message processing system

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030088627A1 (en) * 2001-07-26 2003-05-08 Rothwell Anton C. Intelligent SPAM detection system using an updateable neural analysis engine
US20050257261A1 (en) * 2004-05-02 2005-11-17 Emarkmonitor, Inc. Online fraud solution

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
KIRDA ET AL: 'Protecting Users Against Phishing Attacks with AntiPhish' IEEE PROCEEDINGS OF THE 29TH ANNUAL INTERNATIONAL COMPUTER SOFTWARE AND APPLICATIONS CONFERENCE 2005, pages 527 - 524, XP010837803 *
MERWE ET AL: 'Characteristics and Responsibilities involved in a Phishing Attack' 2005, XP008059482 *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8984289B2 (en) 2003-09-08 2015-03-17 Sonicwall, Inc. Classifying a message based on fraud indicators
GB2422224A (en) * 2005-01-14 2006-07-19 Streamshield Networks Ltd An anti-phishing system for enhancing network security
GB2422224B (en) * 2005-01-14 2007-10-10 Streamshield Networks Ltd An anti-phishing system
JP2007011967A (en) * 2005-07-04 2007-01-18 Nec Corp Mail filtering system, mail filter, and program
US7975297B2 (en) 2005-08-16 2011-07-05 Microsoft Corporation Anti-phishing protection
US9774624B2 (en) 2005-08-16 2017-09-26 Microsoft Technology Licensing, Llc Anti-phishing protection
US9774623B2 (en) 2005-08-16 2017-09-26 Microsoft Technology Licensing, Llc Anti-phishing protection
US10069865B2 (en) 2005-08-16 2018-09-04 Microsoft Technology Licensing, Llc Anti-phishing protection
EP1801745A1 (en) * 2005-12-14 2007-06-27 Aladdin Knowledge Systems, Ltd. Method and system for blocking phishing scams
US8028335B2 (en) 2006-06-19 2011-09-27 Microsoft Corporation Protected environments for protecting users against undesirable activities
EP2017758A1 (en) * 2007-07-02 2009-01-21 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Computer-assisted system and computer-assisted method for content verification

Also Published As

Publication number Publication date
US20070101423A1 (en) 2007-05-03
US8191148B2 (en) 2012-05-29
US20120222111A1 (en) 2012-08-30
US20100095378A1 (en) 2010-04-15
US7665140B2 (en) 2010-02-16
US20080168555A1 (en) 2008-07-10
US8984289B2 (en) 2015-03-17
US7451487B2 (en) 2008-11-11
US8661545B2 (en) 2014-02-25
US20140157409A1 (en) 2014-06-05
WO2005027016A3 (en) 2006-02-23

Similar Documents

Publication Publication Date Title
US8984289B2 (en) Classifying a message based on fraud indicators
US20230344869A1 (en) Detecting phishing attempts
US10834127B1 (en) Detection of business email compromise attacks
Fette et al. Learning to detect phishing emails
US7668921B2 (en) Method and system for phishing detection
US20090089859A1 (en) Method and apparatus for detecting phishing attempts solicited by electronic mail
Ranganayakulu et al. Detecting malicious urls in e-mail–an implementation
US20070094500A1 (en) System and Method for Investigating Phishing Web Sites
US20090300768A1 (en) Method and apparatus for identifying phishing websites in network traffic using generated regular expressions
US20080046970A1 (en) Determining an invalid request
US10341382B2 (en) System and method for filtering electronic messages
Irani et al. Evolutionary study of phishing
CN109039874B (en) Mail auditing method and device based on behavior analysis
Şentürk et al. Email phishing detection and prevention by using data mining techniques
Chawla et al. A survey of phishing attack techniques
WO2017162997A1 (en) A method of protecting a user from messages with links to malicious websites containing homograph attacks
JP4564916B2 (en) Phishing fraud countermeasure method, terminal, server and program
Stone Natural-language processing for intrusion detection
Morovati et al. Detection of Phishing Emails with Email Forensic Analysis and Machine Learning Techniques.
Philomina et al. A comparitative study of machine learning models for the detection of Phishing Websites
US11757816B1 (en) Systems and methods for detecting scam emails
WO2006042480A2 (en) System and method for investigating phishing web sites
Zdziarski et al. Approaches to phishing identification using match and probabilistic digital fingerprinting techniques
Rajpura et al. Enhancement of fake website detection techniques using feature selection and filtering algorithms
Choudhary et al. A Review on Techniques Detection of Phishing URL

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BW BY BZ CA CH CN CO CR CU CZ DK DM DZ EC EE EG ES FI GB GD GE GM HR HU ID IL IN IS JP KE KG KP KZ LC LK LR LS LT LU LV MA MD MK MN MW MX MZ NA NI NO NZ PG PH PL PT RO RU SC SD SE SG SK SY TJ TM TN TR TT TZ UA UG US UZ VN YU ZA ZM

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SZ TZ UG ZM ZW AM AZ BY KG MD RU TJ TM AT BE BG CH CY DE DK EE ES FI FR GB GR HU IE IT MC NL PL PT RO SE SI SK TR BF CF CG CI CM GA GN GQ GW ML MR SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
122 Ep: pct application non-entry in european phase