iocextract 1.16.1

Creator: railscoder56

Last updated:

Add to Cart

Description:

iocextract 1.16.1

iocextract




Indicator of Compromise (IOC) extractor for some of the most commonly ingested artifacts.
Table of contents

Overview

The Problem
Our Solution
Example Use Case


Installation
Usage

Library
Command Line Interface


Helpful Information

FAQ
More Details
Custom Regex
Related Projects
Contributing



Overview
The iocextract package is a library and command line interface (CLI) for extracting URLs, IP addresses, MD5/SHA hashes, email addresses, and YARA rules from text corpora. It allows for you to extract encoded and "defanged" IOCs and optionally decode or refang them.
The Problem
It is common practice for malware analysts or endpoint software to "defang" IOCs such as URLs and IP addresses, in order to prevent accidental exposure to live malicious content. Being able to extract and aggregate these IOCs is often valuable for analysts. Unfortunately, existing "IOC extraction" tools often pass right by them, as they are not caught by standard regex.
For example, the simple defanging technique of surrounding periods with brackets:
127[.]0[.]0[.]1

Existing tools that use a simple IP address regex will ignore this IOC entirely.
Our Solution
By combining specially crafted regex with some custom post-processing, we are able to both detect and deobfuscate "defanged" IOCs. This saves time and effort for the analyst, who might otherwise have to manually find and convert IOCs into machine-readable format.
Example Use Case
Many Twitter users post C2s or other valuable IOC information with defanged URLs.
For example, this tweet from @InQuest:
Recommended reading and great work from @unit42_intel:
https://researchcenter.paloaltonetworks.com/2018/02/unit42-sofacy-attacks-multiple-government-entities/ ...
InQuest customers have had detection for threats delivered from hotfixmsupload[.]com
since 6/3/2017 and cdnverify[.]net since 2/1/18.

If we run this through the extractor, we can easily pull out the URLs:
https://researchcenter.paloaltonetworks.com/2018/02/unit42-sofacy-attacks-multiple-government-entities/
hotfixmsupload[.]com
cdnverify[.]net

Passing in refang=True at extraction time would remove the obfuscation, but since these are real IOCs, let's leave them defanged in our documentation.
Installation
You may need to install the Python development headers in order to install the regex dependency. On Ubuntu/Debian-based systems, try:
sudo apt-get install python-dev

Then install iocextract from pip:
pip install iocextract

If you have problems installing on Windows, try installing regex directly by downloading the appropriate wheel from PyPI and installing via pip:
pip install regex-2018.06.21-cp27-none-win_amd64.whl

Usage
Library
Try extracting some defanged URLs:
import iocextract

content = \
"""
I really love example[.]com!
All the bots are on hxxp://example.com/bad/url these days.
C2: tcp://example[.]com:8989/bad
"""

for url in iocextract.extract_urls(content):
print(url)

# Output

# hxxp://example.com/bad/url
# tcp://example[.]com:8989/bad
# example[.]com
# tcp://example[.]com:8989/bad

NOTE: Some URLs may show up twice if they are caught by multiple regexes.
If you want, you can also "refang", or remove common obfuscation methods from IOCs:
import iocextract

for url in iocextract.extract_urls(content, refang=True):
print(url)

# Output

# http://example.com/bad/url
# http://example.com:8989/bad
# http://example.com
# http://example.com:8989/bad

If you don't want to defang the extracted IOCs at all during extraction, you can disable this as well:
import iocextract

content = \
"""
http://example.com/bad/url
http://example.com:8989/bad
http://example.com
http://example.com:8989/bad
"""

for url in iocextract.extract_urls(content, defang=False):
print(url)

# Output

# http://example.com/bad/url
# http://example.com:8989/bad
# http://example.com
# http://example.com:8989/bad

All extract_* functions in this library return iterators, not lists. The benefit of this behavior is that iocextract can process extremely large inputs, with a very low overhead. However, if for some reason you need to iterate over the IOCs more than once, you will have to save the results as a list:
import iocextract

content = \
"""
I really love example[.]com!
All the bots are on hxxp://example.com/bad/url these days.
C2: tcp://example[.]com:8989/bad
"""

print(list(iocextract.extract_urls(content)))
# ['hxxp://example.com/bad/url', 'tcp://example[.]com:8989/bad', 'example[.]com', 'tcp://example[.]com:8989/bad']

Command Line Interface
A command-line tool is also included:
$ iocextract -h
usage: iocextract [-h] [--input INPUT] [--output OUTPUT] [--extract-emails]
[--extract-ips] [--extract-ipv4s] [--extract-ipv6s]
[--extract-urls] [--extract-yara-rules] [--extract-hashes]
[--custom-regex REGEX_FILE] [--refang] [--strip-urls]
[--wide]

Advanced Indicator of Compromise (IOC) extractor. If no arguments are
specified, the default behavior is to extract all IOCs.

optional arguments:
-h, --help show this help message and exit
--input INPUT default: stdin
--output OUTPUT default: stdout
--extract-emails
--extract-ips
--extract-ipv4s
--extract-ipv6s
--extract-urls
--extract-yara-rules
--extract-hashes
--custom-regex REGEX_FILE file with custom regex strings, one per line, with one capture group each
--refang default: no
--strip-urls remove possible garbage from the end of urls. default: no
--wide preprocess input to allow wide-encoded character matches. default: no

NOTE: Only URLs, emails, and IPv4 addresses can be "refanged".
Helpful Information
FAQ
Are you...

Q. Extracting possibly-defanged IOCs from plain text, like the contents of tweets or blog posts?

A. Yes! This is exactly what iocextract was designed for, and where it performs best. Want to go a step farther and automate extraction and storage? Check out ThreatIngestor.



Q. Extracting URLs that have been hex or base64 encoded?

A. Yes, but the CLI might not give you the best results. Try writing a Python script and calling iocextract.extract_encoded_urls directly.


Note: You will most likely end up with extra garbage at the end of URLs.

Q. Extracting IOCs that have not been defanged, from HTML/XML/RTF?

A. Maybe, but you should consider using the --strip-urls CLI flag (or the strip=True parameter in the library), and you may still get some extra garbage in your output. If you're extracting from HTML, consider using something like Beautiful Soup to first isolate the text content, and then pass that to iocextract, like this.



Q. Extracting IOCs that have not been defanged, from binary data like executables, or very large inputs?

A. There is a very simplistic version of this available when running as a library, but it requires the defang=False parameter and could potentially miss some of the IOCs. The regex in iocextract is designed to be flexible to catch defanged IOCs. If you're unable to collect the information you need, consider using something like Cacador instead.


More Details
This library currently supports the following IOCs:

IP Addresses

IPv4 fully supported
IPv6 partially supported


URLs

With protocol specifier: http, https, tcp, udp, ftp, sftp, ftps
With [.] anchor, even with no protocol specifier
IPv4 and IPv6 (RFC2732) URLs are supported
Hex-encoded URLs with protocol specifier: http, https, ftp
URL-encoded URLs with protocol specifier: http, https, ftp, ftps, sftp
Base64-encoded URLs with protocol specifier: http, https, ftp


Emails

Partially supported, anchoring on @ or at


YARA rules

With imports, includes, and comments


Hashes

MD5
SHA1
SHA256
SHA512


Telephone numbers
Custom regex

With exactly one capture group



For IPv4 addresses, the following defang techniques are supported:



Technique
Defanged
Refanged




. -> [.]
1[.]1[.]1[.]1
1.1.1.1


. -> (.)
1(.)1(.)1(.)1
1.1.1.1


. -> \.
1\.1\.1\.1
1.1.1.1


Partial
1[.1[.1.]1
1.1.1.1


Any combination
1.)1[.1.)1
1.1.1.1



For email addresses, the following defang techniques are supported:



Technique
Defanged
Refanged




. -> [.]
me@example[.]com
me@example.com


. -> (.)
me@example(.)com
me@example.com


. -> {.}
me@example{.}com
me@example.com


. -> _dot_
me@example dot com
me@example.com


@ -> [@]
me[@]example.com
me@example.com


@ -> (@)
me(@)example.com
me@example.com


@ -> {@}
me{@}example.com
me@example.com


@ -> _at_
me at example.com
me@example.com


Partial
me@} example[.com
me@example.com


Added spaces
me@example [.] com
me@example.com


Any combination
me @example [.)com
me@example.com



For URLs, the following defang techniques are supported:



Technique
Defanged
Refanged




. -> [.]
example[.]com/path
http://example.com/path


. -> (.)
example(.)com/path
http://example.com/path


. -> \.
example\.com/path
http://example.com/path


Partial
http://example[.com/path
http://example.com/path


/ -> [/]
http://example.com[/]path
http://example.com/path


Cisco ESA
http:// example .com /path
http://example.com/path


:// -> __
http__example.com/path
http://example.com/path


:// -> :\\
http:\\example.com/path
http://example.com/path


: -> [:]
http[:]//example.com/path
http://example.com/path


hxxp
hxxp://example.com/path
http://example.com/path


Any combination
hxxp__ example( .com[/]path
http://example.com/path


Hex encoded
687474703a2f2f6578616d706c652e636f6d2f70617468
http://example.com/path


URL encoded
http%3A%2F%2fexample%2Ecom%2Fpath
http://example.com/path


Base64 encoded
aHR0cDovL2V4YW1wbGUuY29tL3BhdGgK
http://example.com/path



NOTE: The tables above are not exhaustive, and other URL/defang patterns may also be extracted correctly. If you notice something missing or not working correctly, feel free to let us know via the GitHub Issues.
The base64 regex was generated with @deadpixi's base64 regex tool.
Custom Regex
If you'd like to use the CLI to extract IOCs using your own custom regex, create a plain text file with one regex string per line, and pass it in with the --custom-regex flag. Be sure each regex string includes exactly one capture group.
For example:
http://(example\.com)/
(?:https|ftp)://(example\.com)/

This custom regex file will extrac the domain example.com from matching URLs. The (?: ) noncapture group won't be included in matches.
If you would like to extract the entire match, just put parentheses around your entire regex string, like this:
(https?://.*?.com)

If your regex is invalid, you'll see an error message like this:
Error in custom regex: missing ) at position 5

If your regex does not include a capture group, you'll see an error message like this:
Error in custom regex: no such group

Always use a single capture group when working with custom regex. Here's a quick example:
[
r'(my regex)', # This yields 'my regex' if the pattern matches
r'my (re)gex', # This yields 're' if the pattern matches
]

Using more than a single capture group can cause unexpected results. Check out this example:
[
r'my regex', # This doesn't yield anything
r'(my) (re)gex', # This yields 'my' if the pattern matches
]

Why? Because the result will always yield only the first group match from each regex.
For more complicated regex queries, you can combine capture and non-capture groups like so:
[
r'(?:my|your) (re)gex', # This yields 're' if the pattern matches
]

You can now compare the (?: ) syntax for noncapture groups vs the ( ) syntax for the capture group.
Related Projects
If iocextract doesn't fit your use case, several similar projects exist. Check out the defang and indicators-of-compromise tags on GitHub, as well as:

Cacador in Go
ioc-extractor in JavaScript
Cyobstract in Python

If you'd like to automate IOC extraction, enrichment, export, and more, check out ThreatIngestor.
If you're working with YARA rules, you may be interested in plyara.
Contributing
If you have a defang technique that doesn't make it through the extractor, or if you find any bugs, Pull Requests and Issues are always welcome. The library is released under a GPL-2.0 license.
Who's using iocextract?

InQuest
PacketTotal

Are you using it? Want to see your site listed here? Let us know!

License

For personal and professional use. You cannot resell or redistribute these repositories in their original state.

Customer Reviews

There are no reviews.