Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SOAR-18543] Palo Alto Cortex XDR #3040

Open
wants to merge 8 commits into
base: develop
Choose a base branch
from

Conversation

ablakley-r7
Copy link
Collaborator

Proposed Changes

Description

Describe the proposed changes:

  • Update query start and end time logic to utilise end time of query in non pagination runs. This is to prevent duplicate events being continually processed and raised as new events.
  • Update error handling to return error data in response
  • Update unit tests for new pagination logic
  • Update custom config to read similarly to other plugins

PR Requirements

Developers, verify you have completed the following items by checking them off:

Testing

Unit Tests

Review our documentation on generating and writing plugin unit tests

  • Unit tests written for any new or updated code

In-Product Tests

If you are an InsightConnect customer or have access to an InsightConnect instance, the following in-product tests should be done:

  • Screenshot of job output with the plugin changes
  • Screenshot of the changed connection, actions, or triggers input within the InsightConnect workflow builder

Style

Review the style guide

  • For dependencies, pin OS package and Python package versions
  • For security, set least privileged account with USER nobody in the Dockerfile when possible
  • For size, use the slim SDK images when possible: rapid7/insightconnect-python-3-38-slim-plugin:{sdk-version-num} and rapid7/insightconnect-python-3-38-plugin:{sdk-version-num}
  • For error handling, use of PluginException and ConnectionTestException
  • For logging, use self.logger
  • For docs, use changelog style
  • For docs, validate markdown with insight-plugin validate which calls icon_validate to lint help.md

Functional Checklist

  • Work fully completed
  • Functional
    • Any new actions/triggers include JSON test files in the tests/ directory created with insight-plugin samples
    • Tests should all pass unless it's a negative test. Negative tests have a naming convention of tests/$action_bad.json
    • Unsuccessful tests should fail by raising an exception causing the plugin to die and an object should be returned on successful test
    • Add functioning test results to PR, sanitize any output if necessary
      • Single action/trigger insight-plugin run -T tests/example.json --debug --jq
      • All actions/triggers shortcut insight-plugin run -T all --debug --jq (use PR format at end)
    • Add functioning run results to PR, sanitize any output if necessary
      • Single action/trigger insight-plugin run -R tests/example.json --debug --jq
      • All actions/triggers shortcut insight-plugin run --debug --jq (use PR format at end)

Assessment

You must validate your work to reviewers:

  1. Run insight-plugin validate and make sure everything passes
  2. Run the assessment tool: insight-plugin run -A. For single action validation: insight-plugin run tests/{file}.json -A
  3. Copy (insight-plugin ... | pbcopy) and paste the output in a new post on this PR
  4. Add required screenshots from the In-Product Tests section

@joneill-r7 joneill-r7 requested a review from a team as a code owner January 10, 2025 11:57
@ablakley-r7 ablakley-r7 force-pushed the soar-18543_palo_alto_cortex_xdr branch from 82083d8 to 6c7ef56 Compare January 10, 2025 12:01
@ablakley-r7 ablakley-r7 force-pushed the soar-18543_palo_alto_cortex_xdr branch from dafd15d to d275283 Compare January 13, 2025 08:37
@ablakley-r7 ablakley-r7 force-pushed the soar-18543_palo_alto_cortex_xdr branch from d275283 to a26e0fa Compare January 14, 2025 07:45
self.logger.info("Adjusting start time to cutoff value")
start_time = max_lookback_unix
# Reset search_from and search_to if this is not a backfill
if not custom_config:
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

if we've passed an alert limit into the custom config this could break this logic? is it better to use the lookback type key we've used in other cases?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

You're correct, we should enter this if there is an alert_limit present in the custom config but we're not currently doing a backfill.

@ablakley-r7 ablakley-r7 force-pushed the soar-18543_palo_alto_cortex_xdr branch from cbd1a06 to 8c4a548 Compare January 21, 2025 15:30
Copy link
Collaborator

@joneill-r7 joneill-r7 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It's such a big update so sorry if some of the questions aren't needed! just want to make sure I'm following everything and we're not using resources/time when we don't need to be

default_date_lookback = dt_now - timedelta(days=lookback_days) # if not passed from CPS create on the fly
custom_lookback = custom_config.get(f"max_{LAST_ALERT_TIME}", {})
comparison_date = datetime(**custom_lookback) if custom_lookback else default_date_lookback
now_date_time = self.convert_unix_to_datetime(now_unix)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

should we pass this into get_query_times, we call this as soon as get_query_times gets entered.

)
lookback_values = [bool(custom_timings), custom_config.get("max_lookback_days"), bool(max_lookback_date_time)]
if any(lookback_value for lookback_value in lookback_values):
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we can just do - if any(lookback_values) and it iterates for us?

max_lookback_unix = self.convert_datetime_to_unix(max_lookback_date_time)
if start_time < max_lookback_unix:
self.logger.info(
f"Start time of {self.convert_unix_to_datetime(start_time)} exceeds cutoff of {max_lookback_date_time}"
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

we seem to be doing a lot of converting back and forward. would it be easier to follow keeping this as a datetime obj up until we then decide what start_time we want to keep? and then one conversion of datetime_obj -> unix?

)
self.logger.info("Adjusting start time to cutoff value")
start_time = max_lookback_unix
# Reset search_from and search_to if this is not a backfill
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

if we're changing the state time at all, should we not reset the search_from and search_to?

state[QUERY_START_TIME] = query_values.QUERY_START_TIME
state[QUERY_END_TIME] = query_values.QUERY_END_TIME
state[LAST_SEARCH_FROM] = query_values.QUERY_SEARCH_FROM
state[LAST_SEARCH_TO] = query_values.QUERY_SEARCH_TO
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

do we have an upper limit on this LAST_SEARCH_TO ? want to make sure we don't somehow page so high that we hit an offset limit on their API


# Create a new hash for every new alert
for _, alert in enumerate(alerts):
for alert in alerts:
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

in this loop, should we only add the new hash if it's not deduped? saves the amount of values being saved to the state and also how many are being checked against on the next run?


state[CURRENT_COUNT] = state.get(CURRENT_COUNT, 0) + results_count
new_alerts, new_alert_hashes = self._dedupe_and_get_highest_time(results, state)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I know this isn't changing the logic, but do we think we need to dedupe when we're paging? if we're using offsets we could maybe speed up the task execution time by skipping the dedupe and hashing until we're on the first and last page?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants