For installation instructions, see INSTALL.md. For upgrade instructions, see UPGRADING.md.
Table of Contents:
- Where to get help?
- Configure services
- Configuration
- System Configuration (defaults)
- Pipeline Configuration
- Runtime Configuration
- Harmonization Configuration
- Utilities
- Management
- Error Handling
- Monitoring Logs
- Uninstall
- Integration with ticket systems, etc.
- Frequently Asked Questions
- Additional Information
In case you are lost or something is not discussed in this guide, you might want to subscribe to the intelmq users mailinglist and ask your questions there.
With that clarified, let's dig into the details...
You need to enable and start Redis if not already done. Using systemd it can be done with:
systemctl enable redis.service
systemctl start redis.service
If you installed the packages, standard Linux paths (LSB paths) are used: /var/log/intelmq/
, /etc/intelmq/
, /var/lib/intelmq/
, /var/run/intelmq/
.
Otherwise, the configuration directory is /opt/intelmq/etc/
. Using the environment variable INTELMQ_ROOT_DIR
allows setting any arbitrary root directory.
You can switch this by setting the environment variables INTELMQ_PATHS_NO_OPT
and INTELMQ_PATHS_OPT
, respectively.
- When installing the Python packages, you can set
INTELMQ_PATHS_NO_OPT
to something non-empty to use LSB-paths. - When installing the deb/rpm packages, you can set
INTELMQ_PATHS_OPT
to something non-empty to use/opt/intelmq/
paths, or a path set withINTELMQ_ROOT_DIR
.
The environment variable ROOT_DIR
is meant to set an alternative root directory instead of /
. This is primarily meant for package build environments an analogous to setuptools' --root
parameter. Thus it is only used in LSB-mode.
All configuration files are in the JSON format.
For new installations a default setup with some examples is provided by the intelmqsetup
tool. If this is not the case, make sure the program was run (see installation instructions).
defaults.conf
: default values for all bots and their behavior, e.g. error handling, log options and pipeline configuration. Will be removed in the future.runtime.conf
: Configuration for the individual bots. See Bots for more details.pipeline.conf
: Defines source and destination queues per bot (i.e. where does a bot get its data from, where does it send it to?).BOTS
: Includes configuration hints for all bots. E.g. feed URLs or database connection parameters. Use this as a template forruntime.conf
. This is also read by the intelmq-manager.
To configure a new bot, you need to define and configure it in runtime.conf
using the template from BOTS.
Configure source and destination queues in pipeline.conf
.
Use the IntelMQ Manager mentioned above to generate the configuration files if unsure.
In the shipped examples 4 collectors and parsers, 6 common experts and one output are configured. The default collector and the parser handle data from malware domain list, the file output bot writes all data to /opt/intelmq/var/lib/bots/file-output/events.txt
//var/lib/intelmq/bots/file-output/events.txt
.
All bots inherit this configuration parameters and they can overwrite them using the same parameters in their respective configuration in the ''runtime.conf'' file.
Example:
logging_handler
: Can be one of"file"
or"syslog"
.logging_level
: Defines the system-wide log level that will be use by all bots and the intelmqctl tool. Possible values are:"CRITICAL"
,"ERROR"
,"WARNING"
,"INFO"
and"DEBUG"
.logging_path
: Iflogging_handler
isfile
. Defines the system-wide log-folder that will be use by all bots and the intelmqctl tool. Default value:/opt/intelmq/var/log/
//opt/var/log/intelmq/
.logging_syslog
: Iflogging_handler
issyslog
. Either a list with hostname and UDP port of syslog service, e.g.["localhost", 514]
or a device name/path, e.g. the default"/var/log"
.
We recommend logging_level
WARNING
for production environments and INFO
if you want more details. In any case, watch your free disk space.
You can set these parameters per bot as well. The settings will take effect after the runtime configuration has been (re-)read (which is after loading the defaults configuration. See the intelmqctl documentation).
-
error_log_message
- in case of an error, this option will allow the bot to write the message (report or event) to the log file. Use the following values:true/false
- write or not write message to the log file
-
error_log_exception
- in case of an error, this option will allow the bot to write the error exception to the log file. Use the following values:true/false
- write or not write exception to the log file
-
error_procedure
- in case of an error, this option defines the procedure that the bot will adopt. Use the following values:-
stop
- stop bot after retrying X times (as defined inerror_max_retries
) with a delay between retries (as defined inerror_retry_delay
). If the bot reaches theerror_max_retries
value, it will remove the message from the pipeline and stop. If the optionerror_dump_message
is also enable, the bot will dump the removed message to its dump file (to be found in var/log). -
pass
- will skip this message and will process the next message after retrying X times, removing the current message from pipeline. If the optionerror_dump_message
is also enable, then the bot will dump the removed message to its dump file. After max retries are reached, the rate limit is applied (e.g. a collector bot fetch an unavailable resource does not try forever).
-
-
error_max_retries
- in case of an error, the bot will try to re-start processing the current message X times as defined by this option. int value. -
error_retry_delay
- defines the number of seconds to wait between subsequent re-tries in case of an error. int value. -
error_dump_message
- specifies if the bot will write queued up messages to its dump file (use intelmqdump to re-insert the message).true/false
- write or not write message to the dump file
If the path _on_error
exists for a bot, the message is also sent to this queue, instead of (only) dumping the file if configured to do so.
-
load_balance
- this option allows you to choose the behavior of the queue. Use the following values:true
- splits the messages into several queues without duplicationfalse
- duplicates the messages into each queue- When using AMQP as message broker, take a look at the Multithreading section and the
instances_threads
parameter.
-
broker
- select which broker intelmq can use. Use the following values:redis
- Redis allows some persistence but is not so fast as ZeroMQ (in development). But note that persistence has to be manually activated. See http://redis.io/topics/persistence
-
rate_limit
- time interval (in seconds) between messages processing. int value. -
ssl_ca_certificate
- trusted CA certificate for IMAP connections (supported by some bots). -
source_pipeline_host
- broker IP, FQDN or Unix socket that the bot will use to connect and receive messages. -
source_pipeline_port
- broker port that the bot will use to connect and receive messages. Can be empty for Unix socket. -
source_pipeline_password
- broker password that the bot will use to connect and receive messages. Can be null for unprotected broker. -
source_pipeline_db
- broker database that the bot will use to connect and receive messages (requirement from redis broker). -
destination_pipeline_host
- broker IP, FQDN or Unix socket that the bot will use to connect and send messages. -
destination_pipeline_port
- broker port that the bot will use to connect and send messages. Can be empty for Unix socket. -
destination_pipeline_password
- broker password that the bot will use to connect and send messages. Can be null for unprotected broker. -
destination_pipeline_db
- broker database that the bot will use to connect and send messages (requirement from redis broker). -
http_proxy
- HTTP proxy the that bot will use when performing HTTP requests (e.g. bots/collectors/collector_http.py). The value must follow RFC1738. -
https_proxy
- HTTPS proxy that the bot will use when performing secure HTTPS requests (e.g. bots/collectors/collector_http.py). -
http_user_agent
- user-agent string that the bot will use when performing HTTP/HTTPS requests (e.g. bots/collectors/collector_http.py). -
http_verify_cert
- defines if the bot will verify SSL certificates when performing HTTPS requests (e.g. bots/collectors/collector_http.py).true/false
- verify or not verify SSL certificates
First of all: Do not use it in production environments yet! It has not been tested thoroughly yet.
Supervisor is process manager written in Python. The main advantage is that it take care about processes, so if bot process exit with failure (exit code different than 0), supervisor try to run it again. Another advantage is that it not require writing PID files.
This was tested on Ubuntu 18.04.
Install supervisor. supervisor_twiddler
is extension for supervisor, that makes possible to create process dynamically. (Ubuntu supervisor
package is currently based on Python 2, so supervisor_twiddler
must be installed with Python 2 pip
.)
apt install supervisor python-pip
pip install supervisor_twiddler
Create default config /etc/supervisor/conf.d/intelmq.conf
and restart supervisor
service:
[rpcinterface:twiddler]
supervisor.rpcinterface_factory=supervisor_twiddler.rpcinterface:make_twiddler_rpcinterface
[group:intelmq]
Change IntelMQ process manager in the defaults configuration:
"process_manager": "supervisor",
After this it is possible to manage bots like before with intelmqctl
command.
This configuration is used by each bot to load the source pipeline and destination pipelines associated to each of them. IntelMQ Manager generates this configuration.
Template:
{
...
"<bot ID>": {
"source-queue": "<source pipeline name>",
"destination-queues": [
"<first destination pipeline name>",
"<second destination pipeline name>",
...
]
},
...
}
Note that destination-queues
contains one of the following values:
- None
- string
- list of strings (as in the template above)
- dict of either strings or lists for complex expert bots:
"destination-queues": {
"_default": "<first destination pipeline name>",
"_on_error": "<optional destination pipeline name in case of errors>",
"other-path": [
"<second destination pipeline name>",
"<third destination pipeline name>",
...
],
...
}
In that case, bot will be able to send the message to one of defined paths. The path "_default"
is used if none is not specified.
In case of errors during processing, and the optional path "_on_error"
is specified, the message will be sent to the pipelines given given as on-error.
Other destination queues can be explicitly addressed by the bots, e.g. bots with filtering capabilities.
Example:
{
...
"malware-domain-list-parser": {
"source-queue": "malware-domain-list-parser-queue",
"destination-queues": [
"file-output-queue"
]
},
...
}
Note that a bot must only have one (input) source queue but may have multiple destination queues.
More examples can be found at intelmq/etc/pipeline.conf
directory in IntelMQ repository.
Starting with IntelMQ 1.2 the AMQP protocol is supported as message queue. To use it, install a broker, for example RabbitMQ. The configuration and the differences are outlined here. Keep in mind that it is slower, but has better monitoring capabilities and is more stable. The AMQP support is considered beta, so small problems might occur. So far, only RabbitMQ as broker has been tested.
You can change the broker for single bots (set the parameters in the runtime configuration per bot) or for the whole botnet (in defaults configuration).
You need to set the parameter source_pipeline_broker
/destination_pipeline_broker
to amqp
. There are more parameters available:
destination_pipeline_broker
:"amqp"
destination_pipeline_host
(default:'127.0.0.1'
)destination_pipeline_port
(default: 5672)destination_pipeline_username
destination_pipeline_password
destination_pipeline_socket_timeout
(default: no timeout)destination_pipeline_amqp_exchange
: Only change/set this if you know what you do. If set, the destination queues are not declared as queues, but used as routing key. (default:''
).destination_pipeline_amqp_virtual_host
(default:'/'
)source_pipeline_host
(default:'127.0.0.1'
)source_pipeline_port
(default: 5672)source_pipeline_username
source_pipeline_password
source_pipeline_socket_timeout
(default: no timeout)source_pipeline_amqp_exchange
: Only change/set this if you know what you do. If set, the destination queues are not declared as queues, but used as routing key. (default:''
).source_pipeline_amqp_virtual_host
(default:'/'
)intelmqctl_rabbitmq_monitoring_url
string, see below (default:"http://{host}:15672"
)
For getting the queue sizes, intelmqctl
needs to connect to the monitoring interface of RabbitMQ. If the monitoring interface is not available under "http://{host}:15672" you can manually set using the parameter intelmqctl_rabbitmq_monitoring_url
.
In a RabbitMQ's default configuration you might not provide a user account, as by default the administrator (guest
:guest
) allows full access from localhost. If you create a separate user account, make sure to add the tag "monitoring" to it, otherwise IntelMQ can't fetch the queue sizes.
Setting the statistics (and cache) parameters is necessary when the local redis is running under a non-default host/port. If this is the case, you can set them explicitly:
statistics_database
:3
statistics_host
:"127.0.0.1"
statistics_password
:null
statistics_port
:6379
This configuration is used by each bot to load its specific (runtime) parameters. Usually, the BOTS
file is used to generate runtime.conf
. Also, the IntelMQ Manager generates this configuration. You may edit it manually as well. Be sure to re-load the bot (see the intelmqctl documentation).
Template:
{
"<bot ID>": {
"group": "<bot type (Collector, Parser, Expert, Output)>",
"name": "<human-readable bot name>",
"module": "<bot code (python module)>",
"description": "<generic description of the bot>",
"parameters": {
"<parameter 1>": "<value 1>",
"<parameter 2>": "<value 2>",
"<parameter 3>": "<value 3>"
}
}
}
Example:
{
"malware-domain-list-collector": {
"group": "Collector",
"name": "Malware Domain List",
"module": "intelmq.bots.collectors.http.collector_http",
"description": "Malware Domain List Collector is the bot responsible to get the report from source of information.",
"parameters": {
"http_url": "http://www.malwaredomainlist.com/updatescsv.php",
"feed": "Malware Domain List",
"rate_limit": 3600
}
}
}
More examples can be found in the intelmq/etc/runtime.conf
directory. See Bots for more details.
By default, all of the bots are started when you start the whole botnet, however there is a possibility to disable a bot. This means that the bot will not start every time you start the botnet, but you can start and stop the bot if you specify the bot explicitly. To disable a bot, add the following to your runtime.conf: "enabled": false
. For example:
{
"malware-domain-list-collector": {
"group": "Collector",
"name": "Malware Domain List",
"module": "intelmq.bots.collectors.http.collector_http",
"description": "Malware Domain List Collector is the bot responsible to get the report from source of information.",
"enabled": false,
"parameters": {
"http_url": "http://www.malwaredomainlist.com/updatescsv.php",
"feed": "Malware Domain List",
"rate_limit": 3600
}
}
}
First of all: Do not use it in production environments yet! There are a few bugs, see below
Since IntelMQ 2.0 it is possible to provide the following parameter:
instances_threads
Set it to a non-zero integer, then this number of worker threads will be spawn. This is useful if bots often wait for system resources or if network-based lookups are a bottleneck.
However, there are currently a few cavecats:
- This is not possible for all bots, there are some exceptions (collectors and some outputs), see the FAQ for some reasons.
- Only use it with the AMQP pipeline, as with Redis, messages may get duplicated because there's only one internal queue
- In the logs, you can see the main thread initializing first, then all of the threads which log with the name
[bot-id].[thread-id]
.
This configuration is used to specify the fields for all message types. The harmonization library will load this configuration to check, during the message processing, if the values are compliant to the "harmonization" format. Usually, this configuration doesn't need any change. It is mostly maintained by the intelmq maintainers.
Template:
{
"<message type>": {
"<field 1>": {
"description": "<field 1 description>",
"type": "<field value type>"
},
"<field 2>": {
"description": "<field 2 description>",
"type": "<field value type>"
}
},
}
Example:
{
"event": {
"destination.asn": {
"description": "The autonomous system number from which originated the connection.",
"type": "Integer"
},
"destination.geolocation.cc": {
"description": "Country-Code according to ISO3166-1 alpha-2 for the destination IP.",
"regex": "^[a-zA-Z0-9]{2}$",
"type": "String"
},
},
}
More examples can be found in the intelmq/etc/harmonization.conf
directory.
IntelMQ has a modular structure consisting of bots. There are four types of bots:
- CollectorBots retrieve data from internal or external sources, the output are reports consisting of many individual data sets / log lines.
- ParserBots parse the (report) data by splitting it into individual events (log lines) and giving them a defined structure, see also Data Harmonization for the list of fields an event may be split up into.
- ExpertBots enrich the existing events by e.g. lookup up information such as DNS reverse records, geographic location information (country code) or abuse contacts for an IP address or domain name.
- OutputBots write events to files, databases, (REST)-APIs or any other data sink that you might want to write to.
Each bot has one source queue (except collectors) and can have multiple destination queues (except outputs). But multiple bots can write to the same pipeline (queue), resulting in multiple inputs for the next bot.
Every bot runs in a separate process. A bot is identifiable by a bot id.
Currently only one instance (i.e. with the same bot id) of a bot can run at the same time. Concepts for multiprocessing are being discussed, see this issue: Multiprocessing per queue is not supported #186. Currently you can run multiple processes of the same bot (with different bot ids) in parallel.
Example: multiple gethostbyname bots (with different bot ids) may run in parallel, with the same input queue and sending to the same output queue. Note that the bot providing the input queue must have the load_balance
option set to true
.
IntelMQ has a tool called IntelMQ Manager that gives users an easy way to configure all pipelines with bots that your team needs. For beginners, it's recommended to use the IntelMQ Manager to become acquainted with the functionalities and concepts. The IntelMQ Manager offers some of the possibilities of the intelmqctl tool and has a graphical interface for runtime and pipeline configurations.
See the IntelMQ Manager repository.
Syntax see intelmqctl -h
-
Starting a bot:
intelmqctl start bot-id
-
Stopping a bot:
intelmqctl stop bot-id
-
Reloading a bot:
intelmqctl reload bot-id
-
Restarting a bot:
intelmqctl restart bot-id
-
Get status of a bot:
intelmqctl status bot-id
-
Run a bot directly for debugging purpose and temporarily leverage the logging level to DEBUG:
intelmqctl run bot-id
-
Get a pdb (or ipdb if installed) live console.
intelmqctl run bot-id console
-
See the message that waits in the input queue.
intelmqctl run bot-id message get
-
See additional help for further explanation.
intelmqctl run bot-id --help
-
Starting the botnet (all bots):
intelmqctl start
-
Starting a group of bots:
intelmqctl start --group experts
-
Get a list of all configured bots:
intelmqctl list bots
-
Get a list of all queues:
intelmqctl list queues
If -q is given, only queues with more than one item are listed. -
Get a list of all queues and status of the bots:
intelmqctl list queues-and-status
-
Clear a queue:
intelmqctl clear queue-id
-
Get logs of a bot:
intelmqctl log bot-id number-of-lines log-level
Reads the last lines from bot log. Log level should be one of DEBUG, INFO, ERROR or CRITICAL. Default is INFO. Number of lines defaults to 10, -1 gives all. Result can be longer due to our logging format! -
Upgrade from a previous version:
intelmqctl upgrade-config
Make a backup of your configuration first, also including bot's configuration files.
The "botnet" represents all currently configured bots which are explicitly enabled. It is, in essence, the graph (pipeline.conf) of the bots which are connected together via their input source queues and destination queues.
To get an overview which bots are running, use intelmqctl status
or use the IntelMQ Manager. Set "enabled": true
in the runtime configuration to add a bot to the botnet. By default, bots will be configured as "enabled": true
. See Bots for more details on configuration.
Disabled bots can still be started explicitly using intelmqctl start <bot_id>
, but will remain in the state disabled
if stopped (and not be implicitly enabled by the start
command). They are not started by intelmqctl start
in analogy to the behavior of widely used initialization systems.
In many cases, it is useful to schedule a bot at a specific time (i.e. via cron(1)), for example to collect information from a website every day at midnight. To do this, set run_mode
to scheduled
in the runtime.conf
for the bot. Check out the following example:
"blocklistde-apache-collector": {
"name": "Generic URL Fetcher",
"group": "Collector",
"module": "intelmq.bots.collectors.http.collector_http",
"description": "All IP addresses which have been reported within the last 48 hours as having run attacks on the service Apache, Apache-DDOS, RFI-Attacks.",
"enabled": false,
"run_mode": "scheduled",
"parameters": {
"feed": "Blocklist.de Apache",
"provider": "Blocklist.de",
"http_url": "https://lists.blocklist.de/lists/apache.txt",
"ssl_client_certificate": null
},
},
You can schedule the bot with a crontab-entry like this:
0 0 * * * intelmqctl start blocklistde-apache-collector
Bots configured as scheduled
will exit after the first successful run.
Setting enabled
to false
will cause the bot to not start with intelmqctl start
, but only with an explicit start, in this example intelmqctl start blocklistde-apache-collector
.
Most of the cases, bots will need to be configured as continuous
run mode (the default) in order to have them always running and processing events. Usually, the types of bots that will require the continuous mode will be Parsers, Experts and Outputs. To do this, set run_mode
to continuous
in the runtime.conf
for the bot. Check the following example:
"blocklistde-apache-parser": {
"name": "Blocklist.de Parser",
"group": "Parser",
"module": "intelmq.bots.parsers.blocklistde.parser",
"description": "Blocklist.DE Parser is the bot responsible to parse the report and sanitize the information.",
"enabled": false,
"run_mode": "continuous",
"parameters": {
},
},
You can now start the bot using the following command:
intelmqctl start blocklistde-apache-parser
Bots configured as continuous
will never exit except if there is an error and the error handling configuration requires the bot to exit. See the Error Handling section for more details.
Whilst restart is a mere stop & start, performing intelmqctl reload <bot_id>
will not stop the bot, permitting it to keep the state: the same common behavior as for (Linux) daemons. It will initialize again (including reading all configuration again) after the current action is finished. Also, the rate limit/sleep is continued (with the new time) and not interrupted like with the restart command. So if you have a collector with a rate limit of 24 h, the reload does not trigger a new fetching of the source at the time of the reload, but just 24 h after the last run – with the new configuration.
Which state the bots are keeping depends on the bots of course.
If you are using the default broker (Redis), in some test situations you may need to quickly clear all pipelines and caches. Use the following procedure:
redis-cli FLUSHDB
redis-cli FLUSHALL
When bots are failing due to bad input data or programming errors, they can dump the problematic message to a file along with a traceback, if configured accordingly. These dumps are saved at in the logging directory as [botid].dump
as JSON files. IntelMQ comes with an inspection and reinjection tool, called intelmqdump
. It is an interactive tool to show all dumped files and the number of dumps per file. Choose a file by bot-id or listed numeric id. You can then choose to delete single entries from the file with e 1,3,4
, show a message in more readable format with s 1
(prints the raw-message, can be long!), recover some messages and put them back in the pipeline for the bot by a
or r 0,4,5
. Or delete the file with all dumped messages using d
.
$ intelmqdump -h
usage:
intelmqdump [botid]
intelmqdump [-h|--help]
intelmqdump can inspect dumped messages, show, delete or reinject them into
the pipeline. It's an interactive tool, directly start it to get a list of
available dumps or call it with a known bot id as parameter.
positional arguments:
botid botid to inspect dumps of
optional arguments:
-h, --help show this help message and exit
--truncate TRUNCATE, -t TRUNCATE
Truncate raw-data with more characters than given. 0 for no truncating. Default: 1000.
Interactive actions after a file has been selected:
- r, Recover by IDs
> r id{,id} [queue name]
> r 3,4,6
> r 3,7,90 modify-expert-queue
The messages identified by a consecutive numbering will be stored in the
original queue or the given one and removed from the file.
- a, Recover all
> a [queue name]
> a
> a modify-expert-queue
All messages in the opened file will be recovered to the stored or given
queue and removed from the file.
- e, Delete entries by IDs
> e id{,id}
> e 3,5
The entries will be deleted from the dump file.
- d, Delete file
> d
Delete the opened file as a whole.
- s, Show by IDs
> s id{,id}
> s 0,4,5
Show the selected IP in a readable format. It's still a raw format from
repr, but with newlines for message and traceback.
- v, Edit by ID
> v id
> v 0
> v 1,2
Opens an editor (by calling `sensible-editor`) on the message. The modified message is then saved in the dump.
- q, Quit
> q
$ intelmqdump
id: name (bot id) content
0: alienvault-otx-parser 1 dumps
1: cymru-whois-expert 8 dumps
2: deduplicator-expert 2 dumps
3: dragon-research-group-ssh-parser 2 dumps
4: file-output2 1 dumps
5: fraunhofer-dga-parser 1 dumps
6: spamhaus-cert-parser 4 dumps
7: test-bot 2 dumps
Which dump file to process (id or name)? 3
Processing dragon-research-group-ssh-parser: 2 dumps
0: 2015-09-03T13:13:22.159014 InvalidValue: invalid value u'NA' (<type 'unicode'>) for key u'source.asn'
1: 2015-09-01T14:40:20.973743 InvalidValue: invalid value u'NA' (<type 'unicode'>) for key u'source.asn'
recover (a)ll, delete (e)ntries, (d)elete file, (q)uit, (s)how by ids, (r)ecover by ids? d
Deleted file /opt/intelmq/var/log/dragon-research-group-ssh-parser.dump
Bots and the intelmqdump tool use file locks to prevent writing to already opened files. Bots are trying to lock the file for up to 60 seconds if the dump file is locked already by another process (intelmqdump) and then give up. Intelmqdump does not wait and instead only shows an error message.
By default, the show
command truncates the raw
field of messages at 1000 characters to change this limit or disable truncating at all (value 0), use the --truncate
parameter.
All bots and intelmqctl
log to /opt/intelmq/var/log/
/var/log/intelmq/
(depending on your installation). In case of failures, messages are dumped to the same directory with the file ending .dump
.
tail -f /opt/intelmq/var/log/*.log
tail -f /var/log/intelmq/*.log
If you installed intelmq with native packages: Use the package management tool to remove the package intelmq
. These tools do not remove configuration by default.
If you installed manually via pip (note that this also deletes all configuration and possibly data):
pip3 uninstall intelmq
rm -r /opt/intelmq
First of all, IntelMQ is a message (event) processing system: it collects feeds, processes them, enriches them, filters them and then stores them somewhere or sends them to another system. It does this in a composable, data flow oriented fashion, based on single events. There are no aggregation or grouping features. Now, if you want to integrate IntelMQ with your ticket system or some other system, you need to send its output to somewhere where your ticket system or other services can pick up IntelMQ's data. This could be a database, splunk, or you could send your events directly via email to a ticket system.
Different users came up with different solutions for this, each of them fitting their own organisation. Hence these solutions are not part of the core IntelMQ repository.
- CERT.at uses a postgresql DB (sql output bot) and has a small tool
intelmqcli
which fetches the events in the postgresql DB which are marked as "new" and will group them and send them out via the RT ticket system. - Others, including BSI, use a tool called
intelmq-mailgen
. It sends E-Mails to the recipients, optionally PGP-signed with defined text-templates, CSV formatted attachments with grouped events and generated ticket numbers.
The following lists external github repositories which you might consult for examples on how to integrate IntelMQ into your workflow:
If you came up with another solution for integration, we'd like to hear from you! Please reach out to us on the intelmq-users list.
Consult the FAQ if you encountered any problems.
To enable bash completion on intelmqctl
and intelmqdump
in order to help you run the commands in an easy manner, follow the installation process here.