mirror of
https://github.com/ghndrx/prowler.git
synced 2026-02-10 23:05:05 +00:00
feat(tags): add resource tags to C-D services (#2003)
Co-authored-by: Pepe Fagoaga <pepe@verica.io>
This commit is contained in:
@@ -21,6 +21,7 @@ class cloudformation_outputs_find_secrets(Check):
|
||||
report.region = stack.region
|
||||
report.resource_id = stack.name
|
||||
report.resource_arn = stack.arn
|
||||
report.resource_tags = stack.tags
|
||||
report.status = "PASS"
|
||||
report.status_extended = f"No secrets found in Stack {stack.name} Outputs."
|
||||
if stack.outputs:
|
||||
|
||||
@@ -1,5 +1,7 @@
|
||||
import threading
|
||||
from dataclasses import dataclass
|
||||
from typing import Optional
|
||||
|
||||
from pydantic import BaseModel
|
||||
|
||||
from prowler.lib.logger import logger
|
||||
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
|
||||
@@ -50,6 +52,7 @@ class CloudFormation:
|
||||
Stack(
|
||||
arn=stack["StackId"],
|
||||
name=stack["StackName"],
|
||||
tags=stack.get("Tags"),
|
||||
outputs=outputs,
|
||||
region=regional_client.region,
|
||||
)
|
||||
@@ -82,8 +85,7 @@ class CloudFormation:
|
||||
)
|
||||
|
||||
|
||||
@dataclass
|
||||
class Stack:
|
||||
class Stack(BaseModel):
|
||||
"""Stack holds a CloudFormation Stack"""
|
||||
|
||||
arn: str
|
||||
@@ -92,25 +94,11 @@ class Stack:
|
||||
"""Stacks[].StackName"""
|
||||
outputs: list[str]
|
||||
"""Stacks[].Outputs"""
|
||||
enable_termination_protection: bool
|
||||
enable_termination_protection: bool = False
|
||||
"""Stacks[].EnableTerminationProtection"""
|
||||
root_nested_stack: str
|
||||
root_nested_stack: str = ""
|
||||
"""Stacks[].RootId"""
|
||||
is_nested_stack: str
|
||||
is_nested_stack: bool = False
|
||||
"""True if the Stack is a Nested Stack"""
|
||||
tags: Optional[list] = []
|
||||
region: str
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
arn,
|
||||
name,
|
||||
outputs,
|
||||
region,
|
||||
):
|
||||
self.arn = arn
|
||||
self.name = name
|
||||
self.outputs = outputs
|
||||
self.enable_termination_protection = False
|
||||
self.is_nested_stack = False
|
||||
self.root_nested_stack = ""
|
||||
self.region = region
|
||||
|
||||
@@ -16,6 +16,7 @@ class cloudformation_stacks_termination_protection_enabled(Check):
|
||||
report.region = stack.region
|
||||
report.resource_id = stack.name
|
||||
report.resource_arn = stack.arn
|
||||
report.resource_tags = stack.tags
|
||||
|
||||
if stack.enable_termination_protection:
|
||||
report.status = "PASS"
|
||||
|
||||
@@ -12,6 +12,7 @@ class cloudfront_distributions_field_level_encryption_enabled(Check):
|
||||
report.region = distribution.region
|
||||
report.resource_arn = distribution.arn
|
||||
report.resource_id = distribution.id
|
||||
report.resource_tags = distribution.tags
|
||||
if (
|
||||
distribution.default_cache_config
|
||||
and distribution.default_cache_config.field_level_encryption_id
|
||||
|
||||
@@ -15,6 +15,7 @@ class cloudfront_distributions_geo_restrictions_enabled(Check):
|
||||
report.region = distribution.region
|
||||
report.resource_arn = distribution.arn
|
||||
report.resource_id = distribution.id
|
||||
report.resource_tags = distribution.tags
|
||||
if distribution.geo_restriction_type == GeoRestrictionType.none:
|
||||
report.status = "FAIL"
|
||||
report.status_extended = f"CloudFront Distribution {distribution.id} has Geo restrictions disabled"
|
||||
|
||||
@@ -15,6 +15,7 @@ class cloudfront_distributions_https_enabled(Check):
|
||||
report.region = distribution.region
|
||||
report.resource_arn = distribution.arn
|
||||
report.resource_id = distribution.id
|
||||
report.resource_tags = distribution.tags
|
||||
if (
|
||||
distribution.default_cache_config
|
||||
and distribution.default_cache_config.viewer_protocol_policy
|
||||
|
||||
@@ -12,6 +12,7 @@ class cloudfront_distributions_logging_enabled(Check):
|
||||
report.region = distribution.region
|
||||
report.resource_arn = distribution.arn
|
||||
report.resource_id = distribution.id
|
||||
report.resource_tags = distribution.tags
|
||||
if distribution.logging_enabled or (
|
||||
distribution.default_cache_config
|
||||
and distribution.default_cache_config.realtime_log_config_arn
|
||||
|
||||
@@ -15,6 +15,7 @@ class cloudfront_distributions_using_deprecated_ssl_protocols(Check):
|
||||
report.region = distribution.region
|
||||
report.resource_arn = distribution.arn
|
||||
report.resource_id = distribution.id
|
||||
report.resource_tags = distribution.tags
|
||||
report.status = "PASS"
|
||||
report.status_extended = f"CloudFront Distribution {distribution.id} is not using a deprecated SSL protocol"
|
||||
|
||||
|
||||
@@ -12,6 +12,7 @@ class cloudfront_distributions_using_waf(Check):
|
||||
report.region = distribution.region
|
||||
report.resource_arn = distribution.arn
|
||||
report.resource_id = distribution.id
|
||||
report.resource_tags = distribution.tags
|
||||
if distribution.web_acl_id:
|
||||
report.status = "PASS"
|
||||
report.status_extended = f"CloudFront Distribution {distribution.id} is using AWS WAF web ACL {distribution.web_acl_id}"
|
||||
|
||||
@@ -1,5 +1,8 @@
|
||||
from dataclasses import dataclass
|
||||
from enum import Enum
|
||||
from typing import Optional
|
||||
|
||||
from pydantic import BaseModel
|
||||
|
||||
from prowler.lib.logger import logger
|
||||
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
|
||||
@@ -24,6 +27,9 @@ class CloudFront:
|
||||
self.__get_distribution_config__(
|
||||
self.client, self.distributions, self.region
|
||||
)
|
||||
self.__list_tags_for_resource__(
|
||||
self.client, self.distributions, self.region
|
||||
)
|
||||
|
||||
def __get_session__(self):
|
||||
return self.session
|
||||
@@ -97,6 +103,19 @@ class CloudFront:
|
||||
f"{region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
def __list_tags_for_resource__(self, client, distributions, region):
|
||||
logger.info("CloudFront - List Tags...")
|
||||
try:
|
||||
for distribution in distributions.values():
|
||||
response = client.list_tags_for_resource(Resource=distribution.arn)[
|
||||
"Tags"
|
||||
]
|
||||
distribution.tags = response.get("Items")
|
||||
except Exception as error:
|
||||
logger.error(
|
||||
f"{region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
|
||||
class OriginsSSLProtocols(Enum):
|
||||
SSLv3 = "SSLv3"
|
||||
@@ -128,35 +147,15 @@ class DefaultCacheConfigBehaviour:
|
||||
field_level_encryption_id: str
|
||||
|
||||
|
||||
@dataclass
|
||||
class Distribution:
|
||||
class Distribution(BaseModel):
|
||||
"""Distribution holds a CloudFront Distribution with the required information to run the rela"""
|
||||
|
||||
arn: str
|
||||
id: str
|
||||
region: str
|
||||
logging_enabled: bool
|
||||
default_cache_config: DefaultCacheConfigBehaviour
|
||||
geo_restriction_type: GeoRestrictionType
|
||||
logging_enabled: bool = False
|
||||
default_cache_config: Optional[DefaultCacheConfigBehaviour]
|
||||
geo_restriction_type: Optional[GeoRestrictionType]
|
||||
origins: list
|
||||
web_acl_id: str
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
arn,
|
||||
id,
|
||||
region,
|
||||
origins,
|
||||
logging_enabled=False,
|
||||
default_cache_config=None,
|
||||
geo_restriction_type=None,
|
||||
web_acl_id="",
|
||||
):
|
||||
self.arn = arn
|
||||
self.id = id
|
||||
self.region = region
|
||||
self.logging_enabled = logging_enabled
|
||||
self.default_cache_config = default_cache_config
|
||||
self.geo_restriction_type = geo_restriction_type
|
||||
self.origins = origins
|
||||
self.web_acl_id = web_acl_id
|
||||
web_acl_id: str = ""
|
||||
tags: Optional[list] = []
|
||||
|
||||
@@ -17,6 +17,7 @@ class cloudtrail_cloudwatch_logging_enabled(Check):
|
||||
report.region = trail.region
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
report.status = "PASS"
|
||||
if trail.is_multiregion:
|
||||
report.status_extended = (
|
||||
|
||||
@@ -13,6 +13,7 @@ class cloudtrail_kms_encryption_enabled(Check):
|
||||
report.region = trail.region
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
report.status = "FAIL"
|
||||
if trail.is_multiregion:
|
||||
report.status_extended = (
|
||||
|
||||
@@ -13,6 +13,7 @@ class cloudtrail_log_file_validation_enabled(Check):
|
||||
report.region = trail.region
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
report.status = "FAIL"
|
||||
if trail.is_multiregion:
|
||||
report.status_extended = (
|
||||
|
||||
@@ -16,6 +16,7 @@ class cloudtrail_logs_s3_bucket_access_logging_enabled(Check):
|
||||
report.region = trail.region
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
report.status = "FAIL"
|
||||
if trail.is_multiregion:
|
||||
report.status_extended = f"Multiregion Trail {trail.name} S3 bucket access logging is not enabled for bucket {trail_bucket}"
|
||||
|
||||
@@ -16,6 +16,7 @@ class cloudtrail_logs_s3_bucket_is_not_publicly_accessible(Check):
|
||||
report.region = trail.region
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
report.status = "PASS"
|
||||
if trail.is_multiregion:
|
||||
report.status_extended = f"S3 Bucket {trail_bucket} from multiregion trail {trail.name} is not publicly accessible"
|
||||
|
||||
@@ -16,6 +16,7 @@ class cloudtrail_multi_region_enabled(Check):
|
||||
report.status = "PASS"
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
if trail.is_multiregion:
|
||||
report.status_extended = (
|
||||
f"Trail {trail.name} is multiregion and it is logging"
|
||||
|
||||
@@ -32,6 +32,7 @@ class cloudtrail_s3_dataevents_read_enabled(Check):
|
||||
report.region = trail.region
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
report.status = "PASS"
|
||||
report.status_extended = f"Trail {trail.name} has a classic data event selector to record all S3 object-level API operations."
|
||||
# advanced event selectors
|
||||
@@ -44,6 +45,7 @@ class cloudtrail_s3_dataevents_read_enabled(Check):
|
||||
report.region = trail.region
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
report.status = "PASS"
|
||||
report.status_extended = f"Trail {trail.name} has an advanced data event selector to record all S3 object-level API operations."
|
||||
|
||||
|
||||
@@ -32,6 +32,7 @@ class cloudtrail_s3_dataevents_write_enabled(Check):
|
||||
report.region = trail.region
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
report.status = "PASS"
|
||||
report.status_extended = f"Trail {trail.name} has a classic data event selector to record all S3 object-level API operations."
|
||||
# advanced event selectors
|
||||
@@ -44,6 +45,7 @@ class cloudtrail_s3_dataevents_write_enabled(Check):
|
||||
report.region = trail.region
|
||||
report.resource_id = trail.name
|
||||
report.resource_arn = trail.arn
|
||||
report.resource_tags = trail.tags
|
||||
report.status = "PASS"
|
||||
report.status_extended = f"Trail {trail.name} has an advanced data event selector to record all S3 object-level API operations."
|
||||
findings.append(report)
|
||||
|
||||
@@ -1,5 +1,6 @@
|
||||
import threading
|
||||
from datetime import datetime
|
||||
from typing import Optional
|
||||
|
||||
from pydantic import BaseModel
|
||||
|
||||
@@ -22,6 +23,7 @@ class Cloudtrail:
|
||||
self.__threading_call__(self.__get_trails__)
|
||||
self.__get_trail_status__()
|
||||
self.__get_event_selectors__()
|
||||
self.__list_tags_for_resource__()
|
||||
|
||||
def __get_session__(self):
|
||||
return self.session
|
||||
@@ -132,6 +134,22 @@ class Cloudtrail:
|
||||
f"{client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
def __list_tags_for_resource__(self):
|
||||
logger.info("CloudTrail - List Tags...")
|
||||
try:
|
||||
for trail in self.trails:
|
||||
# Check if trails are in this region
|
||||
if trail.region == trail.home_region:
|
||||
regional_client = self.regional_clients[trail.region]
|
||||
response = regional_client.list_tags(ResourceIdList=[trail.arn])[
|
||||
"ResourceTagList"
|
||||
][0]
|
||||
trail.tags = response.get("TagsList")
|
||||
except Exception as error:
|
||||
logger.error(
|
||||
f"{error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
|
||||
class Event_Selector(BaseModel):
|
||||
is_advanced: bool
|
||||
@@ -151,3 +169,4 @@ class Trail(BaseModel):
|
||||
kms_key: str = None
|
||||
log_group_arn: str = None
|
||||
data_events: list[Event_Selector] = []
|
||||
tags: Optional[list] = []
|
||||
|
||||
@@ -10,6 +10,7 @@ class cloudwatch_log_group_kms_encryption_enabled(Check):
|
||||
report.region = log_group.region
|
||||
report.resource_id = log_group.name
|
||||
report.resource_arn = log_group.arn
|
||||
report.resource_tags = log_group.tags
|
||||
if log_group.kms_id:
|
||||
report.status = "PASS"
|
||||
report.status_extended = f"Log Group {log_group.name} does have AWS KMS key {log_group.kms_id} associated."
|
||||
|
||||
@@ -12,6 +12,7 @@ class cloudwatch_log_group_retention_policy_specific_days_enabled(Check):
|
||||
report.region = log_group.region
|
||||
report.resource_id = log_group.name
|
||||
report.resource_arn = log_group.arn
|
||||
report.resource_tags = log_group.tags
|
||||
if log_group.retention_days < specific_retention_days:
|
||||
report.status = "FAIL"
|
||||
report.status_extended = f"Log Group {log_group.name} has less than {specific_retention_days} days retention period ({log_group.retention_days} days)."
|
||||
|
||||
@@ -1,7 +1,8 @@
|
||||
import threading
|
||||
from dataclasses import dataclass
|
||||
from typing import Optional
|
||||
|
||||
from pydantic import BaseModel
|
||||
|
||||
from prowler.lib.logger import logger
|
||||
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
|
||||
from prowler.providers.aws.aws_provider import generate_regional_clients
|
||||
@@ -22,6 +23,7 @@ class CloudWatch:
|
||||
self.regional_clients = generate_regional_clients(self.service, audit_info)
|
||||
self.metric_alarms = []
|
||||
self.__threading_call__(self.__describe_alarms__)
|
||||
self.__list_tags_for_resource__()
|
||||
|
||||
def __get_session__(self):
|
||||
return self.session
|
||||
@@ -52,11 +54,11 @@ class CloudWatch:
|
||||
namespace = alarm["Namespace"]
|
||||
self.metric_alarms.append(
|
||||
MetricAlarm(
|
||||
alarm["AlarmArn"],
|
||||
alarm["AlarmName"],
|
||||
metric_name,
|
||||
namespace,
|
||||
regional_client.region,
|
||||
arn=alarm["AlarmArn"],
|
||||
name=alarm["AlarmName"],
|
||||
metric=metric_name,
|
||||
name_space=namespace,
|
||||
region=regional_client.region,
|
||||
)
|
||||
)
|
||||
except Exception as error:
|
||||
@@ -64,6 +66,20 @@ class CloudWatch:
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
def __list_tags_for_resource__(self):
|
||||
logger.info("CloudWatch - List Tags...")
|
||||
try:
|
||||
for metric_alarm in self.metric_alarms:
|
||||
regional_client = self.regional_clients[metric_alarm.region]
|
||||
response = regional_client.list_tags_for_resource(
|
||||
ResourceARN=metric_alarm.arn
|
||||
)["Tags"]
|
||||
metric_alarm.tags = response
|
||||
except Exception as error:
|
||||
logger.error(
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
|
||||
################## CloudWatch Logs
|
||||
class Logs:
|
||||
@@ -77,6 +93,7 @@ class Logs:
|
||||
self.log_groups = []
|
||||
self.__threading_call__(self.__describe_metric_filters__)
|
||||
self.__threading_call__(self.__describe_log_groups__)
|
||||
self.__list_tags_for_resource__()
|
||||
|
||||
def __get_session__(self):
|
||||
return self.session
|
||||
@@ -103,11 +120,11 @@ class Logs:
|
||||
):
|
||||
self.metric_filters.append(
|
||||
MetricFilter(
|
||||
filter["filterName"],
|
||||
filter["metricTransformations"][0]["metricName"],
|
||||
filter["filterPattern"],
|
||||
filter["logGroupName"],
|
||||
regional_client.region,
|
||||
name=filter["filterName"],
|
||||
metric=filter["metricTransformations"][0]["metricName"],
|
||||
pattern=filter["filterPattern"],
|
||||
log_group=filter["logGroupName"],
|
||||
region=regional_client.region,
|
||||
)
|
||||
)
|
||||
except Exception as error:
|
||||
@@ -134,11 +151,11 @@ class Logs:
|
||||
retention_days = log_group["retentionInDays"]
|
||||
self.log_groups.append(
|
||||
LogGroup(
|
||||
log_group["arn"],
|
||||
log_group["logGroupName"],
|
||||
retention_days,
|
||||
kms,
|
||||
regional_client.region,
|
||||
arn=log_group["arn"],
|
||||
name=log_group["logGroupName"],
|
||||
retention_days=retention_days,
|
||||
kms_id=kms,
|
||||
region=regional_client.region,
|
||||
)
|
||||
)
|
||||
except Exception as error:
|
||||
@@ -146,71 +163,42 @@ class Logs:
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
def __list_tags_for_resource__(self):
|
||||
logger.info("CloudWatch Logs - List Tags...")
|
||||
try:
|
||||
for log_group in self.log_groups:
|
||||
regional_client = self.regional_clients[log_group.region]
|
||||
response = regional_client.list_tags_for_resource(
|
||||
resourceArn=log_group.arn.replace(":*", "") # Remove the tailing :*
|
||||
)["tags"]
|
||||
log_group.tags = [response]
|
||||
except Exception as error:
|
||||
logger.error(
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
@dataclass
|
||||
class MetricAlarm:
|
||||
|
||||
class MetricAlarm(BaseModel):
|
||||
arn: str
|
||||
name: str
|
||||
metric: Optional[str]
|
||||
name_space: Optional[str]
|
||||
region: str
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
arn,
|
||||
name,
|
||||
metric,
|
||||
name_space,
|
||||
region,
|
||||
):
|
||||
self.arn = arn
|
||||
self.name = name
|
||||
self.metric = metric
|
||||
self.name_space = name_space
|
||||
self.region = region
|
||||
tags: Optional[list] = []
|
||||
|
||||
|
||||
@dataclass
|
||||
class MetricFilter:
|
||||
class MetricFilter(BaseModel):
|
||||
name: str
|
||||
metric: str
|
||||
pattern: str
|
||||
log_group: str
|
||||
region: str
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
name,
|
||||
metric,
|
||||
pattern,
|
||||
log_group,
|
||||
region,
|
||||
):
|
||||
self.name = name
|
||||
self.metric = metric
|
||||
self.pattern = pattern
|
||||
self.log_group = log_group
|
||||
self.region = region
|
||||
|
||||
|
||||
@dataclass
|
||||
class LogGroup:
|
||||
class LogGroup(BaseModel):
|
||||
arn: str
|
||||
name: str
|
||||
retention_days: int
|
||||
kms_id: str
|
||||
kms_id: Optional[str]
|
||||
region: str
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
arn,
|
||||
name,
|
||||
retention_days,
|
||||
kms_id,
|
||||
region,
|
||||
):
|
||||
self.arn = arn
|
||||
self.name = name
|
||||
self.retention_days = retention_days
|
||||
self.kms_id = kms_id
|
||||
self.region = region
|
||||
tags: Optional[list] = []
|
||||
|
||||
@@ -16,6 +16,8 @@ class codeartifact_packages_external_public_publishing_disabled(Check):
|
||||
report = Check_Report_AWS(self.metadata())
|
||||
report.region = repository.region
|
||||
report.resource_id = package.name
|
||||
report.resource_arn = repository.arn
|
||||
report.resource_tags = repository.tags
|
||||
|
||||
if package.latest_version.origin.origin_type in (
|
||||
OriginInformationValues.INTERNAL,
|
||||
|
||||
@@ -21,6 +21,7 @@ class CodeArtifact:
|
||||
self.repositories = {}
|
||||
self.__threading_call__(self.__list_repositories__)
|
||||
self.__threading_call__(self.__list_packages__)
|
||||
self.__list_tags_for_resource__()
|
||||
|
||||
def __get_session__(self):
|
||||
return self.session
|
||||
@@ -148,6 +149,20 @@ class CodeArtifact:
|
||||
f" {error}"
|
||||
)
|
||||
|
||||
def __list_tags_for_resource__(self):
|
||||
logger.info("CodeArtifact - List Tags...")
|
||||
try:
|
||||
for repository in self.repositories.values():
|
||||
regional_client = self.regional_clients[repository.region]
|
||||
response = regional_client.list_tags_for_resource(
|
||||
resourceArn=repository.arn
|
||||
)["tags"]
|
||||
repository.tags = response
|
||||
except Exception as error:
|
||||
logger.error(
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
|
||||
class RestrictionValues(Enum):
|
||||
"""Possible values for the package origin restriction"""
|
||||
@@ -227,3 +242,4 @@ class Repository(BaseModel):
|
||||
domain_owner: str
|
||||
packages: list[Package] = []
|
||||
region: str
|
||||
tags: Optional[list] = []
|
||||
|
||||
@@ -89,9 +89,3 @@ class CodebuildProject:
|
||||
region: str
|
||||
last_invoked_time: Optional[datetime.datetime]
|
||||
buildspec: Optional[str]
|
||||
|
||||
def __init__(self, name, region, last_invoked_time, buildspec):
|
||||
self.name = name
|
||||
self.region = region
|
||||
self.last_invoked_time = last_invoked_time
|
||||
self.buildspec = buildspec
|
||||
|
||||
@@ -1,5 +1,7 @@
|
||||
import threading
|
||||
from dataclasses import dataclass
|
||||
from typing import Optional
|
||||
|
||||
from pydantic import BaseModel
|
||||
|
||||
from prowler.lib.logger import logger
|
||||
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
|
||||
@@ -44,29 +46,29 @@ class Config:
|
||||
if "lastStatus" in recorder:
|
||||
self.recorders.append(
|
||||
Recorder(
|
||||
recorder["name"],
|
||||
recorder["recording"],
|
||||
recorder["lastStatus"],
|
||||
regional_client.region,
|
||||
name=recorder["name"],
|
||||
recording=recorder["recording"],
|
||||
last_status=recorder["lastStatus"],
|
||||
region=regional_client.region,
|
||||
)
|
||||
)
|
||||
else:
|
||||
self.recorders.append(
|
||||
Recorder(
|
||||
recorder["name"],
|
||||
recorder["recording"],
|
||||
None,
|
||||
regional_client.region,
|
||||
name=recorder["name"],
|
||||
recording=recorder["recording"],
|
||||
last_status=None,
|
||||
region=regional_client.region,
|
||||
)
|
||||
)
|
||||
# No config recorders in region
|
||||
if recorders_count == 0:
|
||||
self.recorders.append(
|
||||
Recorder(
|
||||
self.audited_account,
|
||||
None,
|
||||
None,
|
||||
regional_client.region,
|
||||
name=self.audited_account,
|
||||
recording=None,
|
||||
last_status=None,
|
||||
region=regional_client.region,
|
||||
)
|
||||
)
|
||||
|
||||
@@ -76,21 +78,8 @@ class Config:
|
||||
)
|
||||
|
||||
|
||||
@dataclass
|
||||
class Recorder:
|
||||
class Recorder(BaseModel):
|
||||
name: str
|
||||
recording: bool
|
||||
last_status: str
|
||||
recording: Optional[bool]
|
||||
last_status: Optional[str]
|
||||
region: str
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
name,
|
||||
recording,
|
||||
last_status,
|
||||
region,
|
||||
):
|
||||
self.name = name
|
||||
self.recording = recording
|
||||
self.last_status = last_status
|
||||
self.region = region
|
||||
|
||||
@@ -11,6 +11,7 @@ class directoryservice_directory_log_forwarding_enabled(Check):
|
||||
report = Check_Report_AWS(self.metadata())
|
||||
report.region = directory.region
|
||||
report.resource_id = directory.id
|
||||
report.resource_tags = directory.tags
|
||||
if directory.log_subscriptions:
|
||||
report.status = "PASS"
|
||||
report.status_extended = f"Directory Service {directory.id} have log forwarding to CloudWatch enabled"
|
||||
|
||||
@@ -11,6 +11,7 @@ class directoryservice_directory_monitor_notifications(Check):
|
||||
report = Check_Report_AWS(self.metadata())
|
||||
report.region = directory.region
|
||||
report.resource_id = directory.id
|
||||
report.resource_tags = directory.tags
|
||||
if directory.event_topics:
|
||||
report.status = "PASS"
|
||||
report.status_extended = (
|
||||
|
||||
@@ -14,6 +14,7 @@ class directoryservice_directory_snapshots_limit(Check):
|
||||
report = Check_Report_AWS(self.metadata())
|
||||
report.region = directory.region
|
||||
report.resource_id = directory.id
|
||||
report.resource_tags = directory.tags
|
||||
if directory.snapshots_limits:
|
||||
if directory.snapshots_limits.manual_snapshots_limit_reached:
|
||||
report.status = "FAIL"
|
||||
|
||||
@@ -17,6 +17,7 @@ class directoryservice_ldap_certificate_expiration(Check):
|
||||
report = Check_Report_AWS(self.metadata())
|
||||
report.region = directory.region
|
||||
report.resource_id = certificate.id
|
||||
report.resource_tags = directory.tags
|
||||
|
||||
remaining_days_to_expire = (
|
||||
certificate.expiry_date_time
|
||||
|
||||
@@ -15,6 +15,7 @@ class directoryservice_radius_server_security_protocol(Check):
|
||||
report = Check_Report_AWS(self.metadata())
|
||||
report.region = directory.region
|
||||
report.resource_id = directory.id
|
||||
report.resource_tags = directory.tags
|
||||
if (
|
||||
directory.radius_settings.authentication_protocol
|
||||
== AuthenticationProtocol.MS_CHAPv2
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
import threading
|
||||
from datetime import datetime
|
||||
from enum import Enum
|
||||
from typing import Union
|
||||
from typing import Optional, Union
|
||||
|
||||
from pydantic import BaseModel
|
||||
|
||||
@@ -24,6 +24,7 @@ class DirectoryService:
|
||||
self.__threading_call__(self.__describe_event_topics__)
|
||||
self.__threading_call__(self.__list_certificates__)
|
||||
self.__threading_call__(self.__get_snapshot_limits__)
|
||||
self.__list_tags_for_resource__()
|
||||
|
||||
def __get_session__(self):
|
||||
return self.session
|
||||
@@ -199,6 +200,20 @@ class DirectoryService:
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
def __list_tags_for_resource__(self):
|
||||
logger.info("Directory Service - List Tags...")
|
||||
try:
|
||||
for directory in self.directories.values():
|
||||
regional_client = self.regional_clients[directory.region]
|
||||
response = regional_client.list_tags_for_resource(
|
||||
ResourceId=directory.id
|
||||
)["Tags"]
|
||||
directory.tags = response
|
||||
except Exception as error:
|
||||
logger.error(
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
|
||||
class SnapshotLimit(BaseModel):
|
||||
manual_snapshots_limit: int
|
||||
@@ -284,3 +299,4 @@ class Directory(BaseModel):
|
||||
snapshots_limits: SnapshotLimit = None
|
||||
radius_settings: RadiusSettings = None
|
||||
region: str
|
||||
tags: Optional[list] = []
|
||||
|
||||
@@ -15,6 +15,7 @@ class directoryservice_supported_mfa_radius_enabled(Check):
|
||||
report = Check_Report_AWS(self.metadata())
|
||||
report.region = directory.region
|
||||
report.resource_id = directory.id
|
||||
report.resource_tags = directory.tags
|
||||
if directory.radius_settings.status == RadiusStatus.Completed:
|
||||
report.status = "PASS"
|
||||
report.status_extended = (
|
||||
|
||||
@@ -10,6 +10,7 @@ class dynamodb_accelerator_cluster_encryption_enabled(Check):
|
||||
report.resource_id = cluster.name
|
||||
report.resource_arn = cluster.arn
|
||||
report.region = cluster.region
|
||||
report.resource_tags = cluster.tags
|
||||
report.status = "FAIL"
|
||||
report.status_extended = f"DynamoDB cluster {cluster.name} does not have encryption at rest enabled."
|
||||
if cluster.encryption:
|
||||
|
||||
@@ -1,5 +1,7 @@
|
||||
import threading
|
||||
from dataclasses import dataclass
|
||||
from typing import Optional
|
||||
|
||||
from pydantic import BaseModel
|
||||
|
||||
from prowler.lib.logger import logger
|
||||
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
|
||||
@@ -18,6 +20,7 @@ class DynamoDB:
|
||||
self.__threading_call__(self.__list_tables__)
|
||||
self.__describe_table__()
|
||||
self.__describe_continuous_backups__()
|
||||
self.__list_tags_for_resource__()
|
||||
|
||||
def __get_session__(self):
|
||||
return self.session
|
||||
@@ -94,6 +97,20 @@ class DynamoDB:
|
||||
f"{error.__class__.__name__}:{error.__traceback__.tb_lineno} -- {error}"
|
||||
)
|
||||
|
||||
def __list_tags_for_resource__(self):
|
||||
logger.info("DynamoDB - List Tags...")
|
||||
try:
|
||||
for table in self.tables:
|
||||
regional_client = self.regional_clients[table.region]
|
||||
response = regional_client.list_tags_of_resource(ResourceArn=table.arn)[
|
||||
"Tags"
|
||||
]
|
||||
table.tags = response
|
||||
except Exception as error:
|
||||
logger.error(
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
|
||||
################## DynamoDB DAX
|
||||
class DAX:
|
||||
@@ -105,6 +122,7 @@ class DAX:
|
||||
self.regional_clients = generate_regional_clients(self.service, audit_info)
|
||||
self.clusters = []
|
||||
self.__threading_call__(self.__describe_clusters__)
|
||||
self.__list_tags_for_resource__()
|
||||
|
||||
def __get_session__(self):
|
||||
return self.session
|
||||
@@ -137,10 +155,10 @@ class DAX:
|
||||
encryption = True
|
||||
self.clusters.append(
|
||||
Cluster(
|
||||
cluster["ClusterArn"],
|
||||
cluster["ClusterName"],
|
||||
encryption,
|
||||
regional_client.region,
|
||||
arn=cluster["ClusterArn"],
|
||||
name=cluster["ClusterName"],
|
||||
encryption=encryption,
|
||||
region=regional_client.region,
|
||||
)
|
||||
)
|
||||
except Exception as error:
|
||||
@@ -148,47 +166,32 @@ class DAX:
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
def __list_tags_for_resource__(self):
|
||||
logger.info("DAX - List Tags...")
|
||||
try:
|
||||
for cluster in self.clusters:
|
||||
regional_client = self.regional_clients[cluster.region]
|
||||
response = regional_client.list_tags(ResourceName=cluster.name)["Tags"]
|
||||
cluster.tags = response
|
||||
except Exception as error:
|
||||
logger.error(
|
||||
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
|
||||
)
|
||||
|
||||
@dataclass
|
||||
class Table:
|
||||
|
||||
class Table(BaseModel):
|
||||
arn: str
|
||||
name: str
|
||||
encryption_type: str
|
||||
kms_arn: str
|
||||
pitr: bool
|
||||
encryption_type: Optional[str]
|
||||
kms_arn: Optional[str]
|
||||
pitr: bool = False
|
||||
region: str
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
arn,
|
||||
name,
|
||||
encryption_type,
|
||||
kms_arn,
|
||||
region,
|
||||
):
|
||||
self.arn = arn
|
||||
self.name = name
|
||||
self.encryption_type = encryption_type
|
||||
self.kms_arn = kms_arn
|
||||
self.pitr = False
|
||||
self.region = region
|
||||
tags: Optional[list] = []
|
||||
|
||||
|
||||
@dataclass
|
||||
class Cluster:
|
||||
class Cluster(BaseModel):
|
||||
arn: str
|
||||
name: str
|
||||
encryption: str
|
||||
encryption: bool
|
||||
region: str
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
arn,
|
||||
name,
|
||||
encryption,
|
||||
region,
|
||||
):
|
||||
self.arn = arn
|
||||
self.name = name
|
||||
self.encryption = encryption
|
||||
self.region = region
|
||||
tags: Optional[list] = []
|
||||
|
||||
@@ -9,6 +9,7 @@ class dynamodb_tables_kms_cmk_encryption_enabled(Check):
|
||||
report = Check_Report_AWS(self.metadata())
|
||||
report.resource_id = table.name
|
||||
report.resource_arn = table.arn
|
||||
report.resource_tags = table.tags
|
||||
report.region = table.region
|
||||
report.status = "FAIL"
|
||||
report.status_extended = (
|
||||
|
||||
@@ -9,6 +9,7 @@ class dynamodb_tables_pitr_enabled(Check):
|
||||
report = Check_Report_AWS(self.metadata())
|
||||
report.resource_id = table.name
|
||||
report.resource_arn = table.arn
|
||||
report.resource_tags = table.tags
|
||||
report.region = table.region
|
||||
report.status = "FAIL"
|
||||
report.status_extended = f"DynamoDB table {table.name} does not have point-in-time recovery enabled."
|
||||
|
||||
@@ -210,3 +210,7 @@ class Test_CloudFormation_Service:
|
||||
assert cloudformation.stacks[0].is_nested_stack is False
|
||||
assert cloudformation.stacks[0].root_nested_stack == ""
|
||||
assert cloudformation.stacks[0].region == AWS_REGION
|
||||
assert cloudformation.stacks[0].tags == [
|
||||
{"Key": "Tag1", "Value": "Value1"},
|
||||
{"Key": "Tag2", "Value": "Value2"},
|
||||
]
|
||||
|
||||
@@ -31,7 +31,7 @@ class Test_cloudformation_stacks_termination_protection_enabled:
|
||||
Stack(
|
||||
arn="arn:aws:cloudformation:eu-west-1:123456789012:stack/Test-Stack/796c8d26-b390-41d7-a23c-0702c4e78b60",
|
||||
name=stack_name,
|
||||
outputs="",
|
||||
outputs=[],
|
||||
region=AWS_REGION,
|
||||
)
|
||||
]
|
||||
@@ -68,7 +68,7 @@ class Test_cloudformation_stacks_termination_protection_enabled:
|
||||
Stack(
|
||||
arn="arn:aws:cloudformation:eu-west-1:123456789012:stack/Test-Stack/796c8d26-b390-41d7-a23c-0702c4e78b60",
|
||||
name=stack_name,
|
||||
outputs="",
|
||||
outputs=[],
|
||||
region=AWS_REGION,
|
||||
)
|
||||
]
|
||||
|
||||
@@ -133,6 +133,14 @@ def mock_make_api_call(self, operation_name, kwarg):
|
||||
},
|
||||
"ETag": "",
|
||||
}
|
||||
if operation_name == "ListTagsForResource":
|
||||
return {
|
||||
"Tags": {
|
||||
"Items": [
|
||||
{"Key": "test", "Value": "test"},
|
||||
]
|
||||
}
|
||||
}
|
||||
return make_api_call(self, operation_name, kwarg)
|
||||
|
||||
|
||||
@@ -247,3 +255,7 @@ class Test_CloudFront_Service:
|
||||
].default_cache_config.field_level_encryption_id
|
||||
== "enabled"
|
||||
)
|
||||
|
||||
assert cloudfront.distributions[cloudfront_distribution_id].tags == [
|
||||
{"Key": "test", "Value": "test"},
|
||||
]
|
||||
|
||||
@@ -63,7 +63,6 @@ class Test_Cloudtrail_Service:
|
||||
@mock_cloudtrail
|
||||
@mock_s3
|
||||
def test_describe_trails(self):
|
||||
|
||||
cloudtrail_client_us_east_1 = client("cloudtrail", region_name="us-east-1")
|
||||
s3_client_us_east_1 = client("s3", region_name="us-east-1")
|
||||
cloudtrail_client_eu_west_1 = client("cloudtrail", region_name="eu-west-1")
|
||||
@@ -78,10 +77,20 @@ class Test_Cloudtrail_Service:
|
||||
CreateBucketConfiguration={"LocationConstraint": "eu-west-1"},
|
||||
)
|
||||
cloudtrail_client_us_east_1.create_trail(
|
||||
Name=trail_name_us, S3BucketName=bucket_name_us, IsMultiRegionTrail=False
|
||||
Name=trail_name_us,
|
||||
S3BucketName=bucket_name_us,
|
||||
IsMultiRegionTrail=False,
|
||||
TagsList=[
|
||||
{"Key": "test", "Value": "test"},
|
||||
],
|
||||
)
|
||||
cloudtrail_client_eu_west_1.create_trail(
|
||||
Name=trail_name_eu, S3BucketName=bucket_name_eu, IsMultiRegionTrail=False
|
||||
Name=trail_name_eu,
|
||||
S3BucketName=bucket_name_eu,
|
||||
IsMultiRegionTrail=False,
|
||||
TagsList=[
|
||||
{"Key": "test", "Value": "test"},
|
||||
],
|
||||
)
|
||||
audit_info = self.set_mocked_audit_info()
|
||||
cloudtrail = Cloudtrail(audit_info)
|
||||
@@ -101,6 +110,9 @@ class Test_Cloudtrail_Service:
|
||||
trail.s3_bucket == bucket_name_eu
|
||||
or trail.s3_bucket == bucket_name_us
|
||||
)
|
||||
assert trail.tags == [
|
||||
{"Key": "test", "Value": "test"},
|
||||
]
|
||||
|
||||
@mock_cloudtrail
|
||||
@mock_s3
|
||||
|
||||
@@ -72,6 +72,13 @@ def mock_make_api_call(self, operation_name, kwarg):
|
||||
],
|
||||
}
|
||||
|
||||
if operation_name == "ListTagsForResource":
|
||||
return {
|
||||
"tags": [
|
||||
{"key": "test", "value": "test"},
|
||||
]
|
||||
}
|
||||
|
||||
return make_api_call(self, operation_name, kwarg)
|
||||
|
||||
|
||||
@@ -116,6 +123,9 @@ class Test_CodeArtifact_Service:
|
||||
assert codeartifact.repositories
|
||||
assert codeartifact.repositories["test-repository"]
|
||||
assert codeartifact.repositories["test-repository"].name == "test-repository"
|
||||
assert codeartifact.repositories["test-repository"].tags == [
|
||||
{"key": "test", "value": "test"},
|
||||
]
|
||||
assert (
|
||||
codeartifact.repositories["test-repository"].arn
|
||||
== f"arn:aws:codebuild:{AWS_REGION}:{DEFAULT_ACCOUNT_ID}:repository/test-repository"
|
||||
|
||||
@@ -13,6 +13,7 @@ class Test_config_recorder_all_regions_enabled:
|
||||
from prowler.providers.aws.services.config.config_service import Config
|
||||
|
||||
current_audit_info.audited_partition = "aws"
|
||||
current_audit_info.audited_account = "012345678912"
|
||||
current_audit_info.audited_regions = ["eu-west-1", "us-east-1"]
|
||||
|
||||
with mock.patch(
|
||||
@@ -44,7 +45,7 @@ class Test_config_recorder_all_regions_enabled:
|
||||
from prowler.providers.aws.services.config.config_service import Config
|
||||
|
||||
current_audit_info.audited_partition = "aws"
|
||||
current_audit_info.audited_regions = ["eu-west-1", "us-east-1"]
|
||||
current_audit_info.audited_regions = [AWS_REGION]
|
||||
|
||||
with mock.patch(
|
||||
"prowler.providers.aws.services.config.config_recorder_all_regions_enabled.config_recorder_all_regions_enabled.config_client",
|
||||
@@ -57,7 +58,7 @@ class Test_config_recorder_all_regions_enabled:
|
||||
|
||||
check = config_recorder_all_regions_enabled()
|
||||
result = check.execute()
|
||||
assert len(result) == 2
|
||||
assert len(result) == 1
|
||||
# Search for the recorder just created
|
||||
for recorder in result:
|
||||
if recorder.resource_id:
|
||||
@@ -85,7 +86,7 @@ class Test_config_recorder_all_regions_enabled:
|
||||
from prowler.providers.aws.services.config.config_service import Config
|
||||
|
||||
current_audit_info.audited_partition = "aws"
|
||||
current_audit_info.audited_regions = ["eu-west-1", "us-east-1"]
|
||||
current_audit_info.audited_regions = [AWS_REGION]
|
||||
|
||||
with mock.patch(
|
||||
"prowler.providers.aws.services.config.config_recorder_all_regions_enabled.config_recorder_all_regions_enabled.config_client",
|
||||
@@ -98,7 +99,7 @@ class Test_config_recorder_all_regions_enabled:
|
||||
|
||||
check = config_recorder_all_regions_enabled()
|
||||
result = check.execute()
|
||||
assert len(result) == 2
|
||||
assert len(result) == 1
|
||||
# Search for the recorder just created
|
||||
for recorder in result:
|
||||
if recorder.resource_id:
|
||||
|
||||
@@ -92,7 +92,12 @@ def mock_make_api_call(self, operation_name, kwarg):
|
||||
"ManualSnapshotsLimitReached": True,
|
||||
}
|
||||
}
|
||||
|
||||
if operation_name == "ListTagsForResource":
|
||||
return {
|
||||
"Tags": [
|
||||
{"Key": "string", "Value": "string"},
|
||||
],
|
||||
}
|
||||
return make_api_call(self, operation_name, kwarg)
|
||||
|
||||
|
||||
@@ -145,6 +150,9 @@ class Test_DirectoryService_Service:
|
||||
)
|
||||
assert directoryservice.directories["d-12345a1b2"].name == "test-directory"
|
||||
assert directoryservice.directories["d-12345a1b2"].region == AWS_REGION
|
||||
assert directoryservice.directories["d-12345a1b2"].tags == [
|
||||
{"Key": "string", "Value": "string"},
|
||||
]
|
||||
assert (
|
||||
directoryservice.directories[
|
||||
"d-12345a1b2"
|
||||
|
||||
@@ -121,6 +121,9 @@ class Test_DynamoDB_Service:
|
||||
{"AttributeName": "app", "KeyType": "RANGE"},
|
||||
],
|
||||
BillingMode="PAY_PER_REQUEST",
|
||||
Tags=[
|
||||
{"Key": "test", "Value": "test"},
|
||||
],
|
||||
)["TableDescription"]
|
||||
# DynamoDB client for this test class
|
||||
audit_info = self.set_mocked_audit_info()
|
||||
@@ -129,6 +132,9 @@ class Test_DynamoDB_Service:
|
||||
assert dynamo.tables[0].arn == table["TableArn"]
|
||||
assert dynamo.tables[0].name == "test1"
|
||||
assert dynamo.tables[0].region == AWS_REGION
|
||||
assert dynamo.tables[0].tags == [
|
||||
{"Key": "test", "Value": "test"},
|
||||
]
|
||||
|
||||
# Test DynamoDB Describe Table
|
||||
@mock_dynamodb
|
||||
@@ -174,6 +180,9 @@ class Test_DynamoDB_Service:
|
||||
ReplicationFactor=3,
|
||||
IamRoleArn=iam_role_arn,
|
||||
SSESpecification={"Enabled": True},
|
||||
Tags=[
|
||||
{"Key": "test", "Value": "test"},
|
||||
],
|
||||
)
|
||||
dax_client.create_cluster(
|
||||
ClusterName="daxcluster2",
|
||||
@@ -181,6 +190,9 @@ class Test_DynamoDB_Service:
|
||||
ReplicationFactor=3,
|
||||
IamRoleArn=iam_role_arn,
|
||||
SSESpecification={"Enabled": True},
|
||||
Tags=[
|
||||
{"Key": "test", "Value": "test"},
|
||||
],
|
||||
)
|
||||
# DAX client for this test class
|
||||
audit_info = self.set_mocked_audit_info()
|
||||
@@ -190,3 +202,9 @@ class Test_DynamoDB_Service:
|
||||
assert dax.clusters[1].name == "daxcluster2"
|
||||
assert dax.clusters[0].region == AWS_REGION
|
||||
assert dax.clusters[1].region == AWS_REGION
|
||||
assert dax.clusters[0].tags == [
|
||||
{"Key": "test", "Value": "test"},
|
||||
]
|
||||
assert dax.clusters[1].tags == [
|
||||
{"Key": "test", "Value": "test"},
|
||||
]
|
||||
|
||||
Reference in New Issue
Block a user