feat(tags): add resource tags to C-D services (#2003)

Co-authored-by: Pepe Fagoaga <pepe@verica.io>
This commit is contained in:
Sergio Garcia
2023-03-02 13:14:53 +01:00
committed by GitHub
parent 024ffb1117
commit da834c0935
45 changed files with 306 additions and 198 deletions

View File

@@ -21,6 +21,7 @@ class cloudformation_outputs_find_secrets(Check):
report.region = stack.region
report.resource_id = stack.name
report.resource_arn = stack.arn
report.resource_tags = stack.tags
report.status = "PASS"
report.status_extended = f"No secrets found in Stack {stack.name} Outputs."
if stack.outputs:

View File

@@ -1,5 +1,7 @@
import threading
from dataclasses import dataclass
from typing import Optional
from pydantic import BaseModel
from prowler.lib.logger import logger
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
@@ -50,6 +52,7 @@ class CloudFormation:
Stack(
arn=stack["StackId"],
name=stack["StackName"],
tags=stack.get("Tags"),
outputs=outputs,
region=regional_client.region,
)
@@ -82,8 +85,7 @@ class CloudFormation:
)
@dataclass
class Stack:
class Stack(BaseModel):
"""Stack holds a CloudFormation Stack"""
arn: str
@@ -92,25 +94,11 @@ class Stack:
"""Stacks[].StackName"""
outputs: list[str]
"""Stacks[].Outputs"""
enable_termination_protection: bool
enable_termination_protection: bool = False
"""Stacks[].EnableTerminationProtection"""
root_nested_stack: str
root_nested_stack: str = ""
"""Stacks[].RootId"""
is_nested_stack: str
is_nested_stack: bool = False
"""True if the Stack is a Nested Stack"""
tags: Optional[list] = []
region: str
def __init__(
self,
arn,
name,
outputs,
region,
):
self.arn = arn
self.name = name
self.outputs = outputs
self.enable_termination_protection = False
self.is_nested_stack = False
self.root_nested_stack = ""
self.region = region

View File

@@ -16,6 +16,7 @@ class cloudformation_stacks_termination_protection_enabled(Check):
report.region = stack.region
report.resource_id = stack.name
report.resource_arn = stack.arn
report.resource_tags = stack.tags
if stack.enable_termination_protection:
report.status = "PASS"

View File

@@ -12,6 +12,7 @@ class cloudfront_distributions_field_level_encryption_enabled(Check):
report.region = distribution.region
report.resource_arn = distribution.arn
report.resource_id = distribution.id
report.resource_tags = distribution.tags
if (
distribution.default_cache_config
and distribution.default_cache_config.field_level_encryption_id

View File

@@ -15,6 +15,7 @@ class cloudfront_distributions_geo_restrictions_enabled(Check):
report.region = distribution.region
report.resource_arn = distribution.arn
report.resource_id = distribution.id
report.resource_tags = distribution.tags
if distribution.geo_restriction_type == GeoRestrictionType.none:
report.status = "FAIL"
report.status_extended = f"CloudFront Distribution {distribution.id} has Geo restrictions disabled"

View File

@@ -15,6 +15,7 @@ class cloudfront_distributions_https_enabled(Check):
report.region = distribution.region
report.resource_arn = distribution.arn
report.resource_id = distribution.id
report.resource_tags = distribution.tags
if (
distribution.default_cache_config
and distribution.default_cache_config.viewer_protocol_policy

View File

@@ -12,6 +12,7 @@ class cloudfront_distributions_logging_enabled(Check):
report.region = distribution.region
report.resource_arn = distribution.arn
report.resource_id = distribution.id
report.resource_tags = distribution.tags
if distribution.logging_enabled or (
distribution.default_cache_config
and distribution.default_cache_config.realtime_log_config_arn

View File

@@ -15,6 +15,7 @@ class cloudfront_distributions_using_deprecated_ssl_protocols(Check):
report.region = distribution.region
report.resource_arn = distribution.arn
report.resource_id = distribution.id
report.resource_tags = distribution.tags
report.status = "PASS"
report.status_extended = f"CloudFront Distribution {distribution.id} is not using a deprecated SSL protocol"

View File

@@ -12,6 +12,7 @@ class cloudfront_distributions_using_waf(Check):
report.region = distribution.region
report.resource_arn = distribution.arn
report.resource_id = distribution.id
report.resource_tags = distribution.tags
if distribution.web_acl_id:
report.status = "PASS"
report.status_extended = f"CloudFront Distribution {distribution.id} is using AWS WAF web ACL {distribution.web_acl_id}"

View File

@@ -1,5 +1,8 @@
from dataclasses import dataclass
from enum import Enum
from typing import Optional
from pydantic import BaseModel
from prowler.lib.logger import logger
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
@@ -24,6 +27,9 @@ class CloudFront:
self.__get_distribution_config__(
self.client, self.distributions, self.region
)
self.__list_tags_for_resource__(
self.client, self.distributions, self.region
)
def __get_session__(self):
return self.session
@@ -97,6 +103,19 @@ class CloudFront:
f"{region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
def __list_tags_for_resource__(self, client, distributions, region):
logger.info("CloudFront - List Tags...")
try:
for distribution in distributions.values():
response = client.list_tags_for_resource(Resource=distribution.arn)[
"Tags"
]
distribution.tags = response.get("Items")
except Exception as error:
logger.error(
f"{region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
class OriginsSSLProtocols(Enum):
SSLv3 = "SSLv3"
@@ -128,35 +147,15 @@ class DefaultCacheConfigBehaviour:
field_level_encryption_id: str
@dataclass
class Distribution:
class Distribution(BaseModel):
"""Distribution holds a CloudFront Distribution with the required information to run the rela"""
arn: str
id: str
region: str
logging_enabled: bool
default_cache_config: DefaultCacheConfigBehaviour
geo_restriction_type: GeoRestrictionType
logging_enabled: bool = False
default_cache_config: Optional[DefaultCacheConfigBehaviour]
geo_restriction_type: Optional[GeoRestrictionType]
origins: list
web_acl_id: str
def __init__(
self,
arn,
id,
region,
origins,
logging_enabled=False,
default_cache_config=None,
geo_restriction_type=None,
web_acl_id="",
):
self.arn = arn
self.id = id
self.region = region
self.logging_enabled = logging_enabled
self.default_cache_config = default_cache_config
self.geo_restriction_type = geo_restriction_type
self.origins = origins
self.web_acl_id = web_acl_id
web_acl_id: str = ""
tags: Optional[list] = []

View File

@@ -17,6 +17,7 @@ class cloudtrail_cloudwatch_logging_enabled(Check):
report.region = trail.region
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
report.status = "PASS"
if trail.is_multiregion:
report.status_extended = (

View File

@@ -13,6 +13,7 @@ class cloudtrail_kms_encryption_enabled(Check):
report.region = trail.region
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
report.status = "FAIL"
if trail.is_multiregion:
report.status_extended = (

View File

@@ -13,6 +13,7 @@ class cloudtrail_log_file_validation_enabled(Check):
report.region = trail.region
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
report.status = "FAIL"
if trail.is_multiregion:
report.status_extended = (

View File

@@ -16,6 +16,7 @@ class cloudtrail_logs_s3_bucket_access_logging_enabled(Check):
report.region = trail.region
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
report.status = "FAIL"
if trail.is_multiregion:
report.status_extended = f"Multiregion Trail {trail.name} S3 bucket access logging is not enabled for bucket {trail_bucket}"

View File

@@ -16,6 +16,7 @@ class cloudtrail_logs_s3_bucket_is_not_publicly_accessible(Check):
report.region = trail.region
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
report.status = "PASS"
if trail.is_multiregion:
report.status_extended = f"S3 Bucket {trail_bucket} from multiregion trail {trail.name} is not publicly accessible"

View File

@@ -16,6 +16,7 @@ class cloudtrail_multi_region_enabled(Check):
report.status = "PASS"
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
if trail.is_multiregion:
report.status_extended = (
f"Trail {trail.name} is multiregion and it is logging"

View File

@@ -32,6 +32,7 @@ class cloudtrail_s3_dataevents_read_enabled(Check):
report.region = trail.region
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
report.status = "PASS"
report.status_extended = f"Trail {trail.name} has a classic data event selector to record all S3 object-level API operations."
# advanced event selectors
@@ -44,6 +45,7 @@ class cloudtrail_s3_dataevents_read_enabled(Check):
report.region = trail.region
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
report.status = "PASS"
report.status_extended = f"Trail {trail.name} has an advanced data event selector to record all S3 object-level API operations."

View File

@@ -32,6 +32,7 @@ class cloudtrail_s3_dataevents_write_enabled(Check):
report.region = trail.region
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
report.status = "PASS"
report.status_extended = f"Trail {trail.name} has a classic data event selector to record all S3 object-level API operations."
# advanced event selectors
@@ -44,6 +45,7 @@ class cloudtrail_s3_dataevents_write_enabled(Check):
report.region = trail.region
report.resource_id = trail.name
report.resource_arn = trail.arn
report.resource_tags = trail.tags
report.status = "PASS"
report.status_extended = f"Trail {trail.name} has an advanced data event selector to record all S3 object-level API operations."
findings.append(report)

View File

@@ -1,5 +1,6 @@
import threading
from datetime import datetime
from typing import Optional
from pydantic import BaseModel
@@ -22,6 +23,7 @@ class Cloudtrail:
self.__threading_call__(self.__get_trails__)
self.__get_trail_status__()
self.__get_event_selectors__()
self.__list_tags_for_resource__()
def __get_session__(self):
return self.session
@@ -132,6 +134,22 @@ class Cloudtrail:
f"{client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
def __list_tags_for_resource__(self):
logger.info("CloudTrail - List Tags...")
try:
for trail in self.trails:
# Check if trails are in this region
if trail.region == trail.home_region:
regional_client = self.regional_clients[trail.region]
response = regional_client.list_tags(ResourceIdList=[trail.arn])[
"ResourceTagList"
][0]
trail.tags = response.get("TagsList")
except Exception as error:
logger.error(
f"{error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
class Event_Selector(BaseModel):
is_advanced: bool
@@ -151,3 +169,4 @@ class Trail(BaseModel):
kms_key: str = None
log_group_arn: str = None
data_events: list[Event_Selector] = []
tags: Optional[list] = []

View File

@@ -10,6 +10,7 @@ class cloudwatch_log_group_kms_encryption_enabled(Check):
report.region = log_group.region
report.resource_id = log_group.name
report.resource_arn = log_group.arn
report.resource_tags = log_group.tags
if log_group.kms_id:
report.status = "PASS"
report.status_extended = f"Log Group {log_group.name} does have AWS KMS key {log_group.kms_id} associated."

View File

@@ -12,6 +12,7 @@ class cloudwatch_log_group_retention_policy_specific_days_enabled(Check):
report.region = log_group.region
report.resource_id = log_group.name
report.resource_arn = log_group.arn
report.resource_tags = log_group.tags
if log_group.retention_days < specific_retention_days:
report.status = "FAIL"
report.status_extended = f"Log Group {log_group.name} has less than {specific_retention_days} days retention period ({log_group.retention_days} days)."

View File

@@ -1,7 +1,8 @@
import threading
from dataclasses import dataclass
from typing import Optional
from pydantic import BaseModel
from prowler.lib.logger import logger
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
from prowler.providers.aws.aws_provider import generate_regional_clients
@@ -22,6 +23,7 @@ class CloudWatch:
self.regional_clients = generate_regional_clients(self.service, audit_info)
self.metric_alarms = []
self.__threading_call__(self.__describe_alarms__)
self.__list_tags_for_resource__()
def __get_session__(self):
return self.session
@@ -52,11 +54,11 @@ class CloudWatch:
namespace = alarm["Namespace"]
self.metric_alarms.append(
MetricAlarm(
alarm["AlarmArn"],
alarm["AlarmName"],
metric_name,
namespace,
regional_client.region,
arn=alarm["AlarmArn"],
name=alarm["AlarmName"],
metric=metric_name,
name_space=namespace,
region=regional_client.region,
)
)
except Exception as error:
@@ -64,6 +66,20 @@ class CloudWatch:
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
def __list_tags_for_resource__(self):
logger.info("CloudWatch - List Tags...")
try:
for metric_alarm in self.metric_alarms:
regional_client = self.regional_clients[metric_alarm.region]
response = regional_client.list_tags_for_resource(
ResourceARN=metric_alarm.arn
)["Tags"]
metric_alarm.tags = response
except Exception as error:
logger.error(
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
################## CloudWatch Logs
class Logs:
@@ -77,6 +93,7 @@ class Logs:
self.log_groups = []
self.__threading_call__(self.__describe_metric_filters__)
self.__threading_call__(self.__describe_log_groups__)
self.__list_tags_for_resource__()
def __get_session__(self):
return self.session
@@ -103,11 +120,11 @@ class Logs:
):
self.metric_filters.append(
MetricFilter(
filter["filterName"],
filter["metricTransformations"][0]["metricName"],
filter["filterPattern"],
filter["logGroupName"],
regional_client.region,
name=filter["filterName"],
metric=filter["metricTransformations"][0]["metricName"],
pattern=filter["filterPattern"],
log_group=filter["logGroupName"],
region=regional_client.region,
)
)
except Exception as error:
@@ -134,11 +151,11 @@ class Logs:
retention_days = log_group["retentionInDays"]
self.log_groups.append(
LogGroup(
log_group["arn"],
log_group["logGroupName"],
retention_days,
kms,
regional_client.region,
arn=log_group["arn"],
name=log_group["logGroupName"],
retention_days=retention_days,
kms_id=kms,
region=regional_client.region,
)
)
except Exception as error:
@@ -146,71 +163,42 @@ class Logs:
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
def __list_tags_for_resource__(self):
logger.info("CloudWatch Logs - List Tags...")
try:
for log_group in self.log_groups:
regional_client = self.regional_clients[log_group.region]
response = regional_client.list_tags_for_resource(
resourceArn=log_group.arn.replace(":*", "") # Remove the tailing :*
)["tags"]
log_group.tags = [response]
except Exception as error:
logger.error(
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
@dataclass
class MetricAlarm:
class MetricAlarm(BaseModel):
arn: str
name: str
metric: Optional[str]
name_space: Optional[str]
region: str
def __init__(
self,
arn,
name,
metric,
name_space,
region,
):
self.arn = arn
self.name = name
self.metric = metric
self.name_space = name_space
self.region = region
tags: Optional[list] = []
@dataclass
class MetricFilter:
class MetricFilter(BaseModel):
name: str
metric: str
pattern: str
log_group: str
region: str
def __init__(
self,
name,
metric,
pattern,
log_group,
region,
):
self.name = name
self.metric = metric
self.pattern = pattern
self.log_group = log_group
self.region = region
@dataclass
class LogGroup:
class LogGroup(BaseModel):
arn: str
name: str
retention_days: int
kms_id: str
kms_id: Optional[str]
region: str
def __init__(
self,
arn,
name,
retention_days,
kms_id,
region,
):
self.arn = arn
self.name = name
self.retention_days = retention_days
self.kms_id = kms_id
self.region = region
tags: Optional[list] = []

View File

@@ -16,6 +16,8 @@ class codeartifact_packages_external_public_publishing_disabled(Check):
report = Check_Report_AWS(self.metadata())
report.region = repository.region
report.resource_id = package.name
report.resource_arn = repository.arn
report.resource_tags = repository.tags
if package.latest_version.origin.origin_type in (
OriginInformationValues.INTERNAL,

View File

@@ -21,6 +21,7 @@ class CodeArtifact:
self.repositories = {}
self.__threading_call__(self.__list_repositories__)
self.__threading_call__(self.__list_packages__)
self.__list_tags_for_resource__()
def __get_session__(self):
return self.session
@@ -148,6 +149,20 @@ class CodeArtifact:
f" {error}"
)
def __list_tags_for_resource__(self):
logger.info("CodeArtifact - List Tags...")
try:
for repository in self.repositories.values():
regional_client = self.regional_clients[repository.region]
response = regional_client.list_tags_for_resource(
resourceArn=repository.arn
)["tags"]
repository.tags = response
except Exception as error:
logger.error(
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
class RestrictionValues(Enum):
"""Possible values for the package origin restriction"""
@@ -227,3 +242,4 @@ class Repository(BaseModel):
domain_owner: str
packages: list[Package] = []
region: str
tags: Optional[list] = []

View File

@@ -89,9 +89,3 @@ class CodebuildProject:
region: str
last_invoked_time: Optional[datetime.datetime]
buildspec: Optional[str]
def __init__(self, name, region, last_invoked_time, buildspec):
self.name = name
self.region = region
self.last_invoked_time = last_invoked_time
self.buildspec = buildspec

View File

@@ -1,5 +1,7 @@
import threading
from dataclasses import dataclass
from typing import Optional
from pydantic import BaseModel
from prowler.lib.logger import logger
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
@@ -44,29 +46,29 @@ class Config:
if "lastStatus" in recorder:
self.recorders.append(
Recorder(
recorder["name"],
recorder["recording"],
recorder["lastStatus"],
regional_client.region,
name=recorder["name"],
recording=recorder["recording"],
last_status=recorder["lastStatus"],
region=regional_client.region,
)
)
else:
self.recorders.append(
Recorder(
recorder["name"],
recorder["recording"],
None,
regional_client.region,
name=recorder["name"],
recording=recorder["recording"],
last_status=None,
region=regional_client.region,
)
)
# No config recorders in region
if recorders_count == 0:
self.recorders.append(
Recorder(
self.audited_account,
None,
None,
regional_client.region,
name=self.audited_account,
recording=None,
last_status=None,
region=regional_client.region,
)
)
@@ -76,21 +78,8 @@ class Config:
)
@dataclass
class Recorder:
class Recorder(BaseModel):
name: str
recording: bool
last_status: str
recording: Optional[bool]
last_status: Optional[str]
region: str
def __init__(
self,
name,
recording,
last_status,
region,
):
self.name = name
self.recording = recording
self.last_status = last_status
self.region = region

View File

@@ -11,6 +11,7 @@ class directoryservice_directory_log_forwarding_enabled(Check):
report = Check_Report_AWS(self.metadata())
report.region = directory.region
report.resource_id = directory.id
report.resource_tags = directory.tags
if directory.log_subscriptions:
report.status = "PASS"
report.status_extended = f"Directory Service {directory.id} have log forwarding to CloudWatch enabled"

View File

@@ -11,6 +11,7 @@ class directoryservice_directory_monitor_notifications(Check):
report = Check_Report_AWS(self.metadata())
report.region = directory.region
report.resource_id = directory.id
report.resource_tags = directory.tags
if directory.event_topics:
report.status = "PASS"
report.status_extended = (

View File

@@ -14,6 +14,7 @@ class directoryservice_directory_snapshots_limit(Check):
report = Check_Report_AWS(self.metadata())
report.region = directory.region
report.resource_id = directory.id
report.resource_tags = directory.tags
if directory.snapshots_limits:
if directory.snapshots_limits.manual_snapshots_limit_reached:
report.status = "FAIL"

View File

@@ -17,6 +17,7 @@ class directoryservice_ldap_certificate_expiration(Check):
report = Check_Report_AWS(self.metadata())
report.region = directory.region
report.resource_id = certificate.id
report.resource_tags = directory.tags
remaining_days_to_expire = (
certificate.expiry_date_time

View File

@@ -15,6 +15,7 @@ class directoryservice_radius_server_security_protocol(Check):
report = Check_Report_AWS(self.metadata())
report.region = directory.region
report.resource_id = directory.id
report.resource_tags = directory.tags
if (
directory.radius_settings.authentication_protocol
== AuthenticationProtocol.MS_CHAPv2

View File

@@ -1,7 +1,7 @@
import threading
from datetime import datetime
from enum import Enum
from typing import Union
from typing import Optional, Union
from pydantic import BaseModel
@@ -24,6 +24,7 @@ class DirectoryService:
self.__threading_call__(self.__describe_event_topics__)
self.__threading_call__(self.__list_certificates__)
self.__threading_call__(self.__get_snapshot_limits__)
self.__list_tags_for_resource__()
def __get_session__(self):
return self.session
@@ -199,6 +200,20 @@ class DirectoryService:
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
def __list_tags_for_resource__(self):
logger.info("Directory Service - List Tags...")
try:
for directory in self.directories.values():
regional_client = self.regional_clients[directory.region]
response = regional_client.list_tags_for_resource(
ResourceId=directory.id
)["Tags"]
directory.tags = response
except Exception as error:
logger.error(
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
class SnapshotLimit(BaseModel):
manual_snapshots_limit: int
@@ -284,3 +299,4 @@ class Directory(BaseModel):
snapshots_limits: SnapshotLimit = None
radius_settings: RadiusSettings = None
region: str
tags: Optional[list] = []

View File

@@ -15,6 +15,7 @@ class directoryservice_supported_mfa_radius_enabled(Check):
report = Check_Report_AWS(self.metadata())
report.region = directory.region
report.resource_id = directory.id
report.resource_tags = directory.tags
if directory.radius_settings.status == RadiusStatus.Completed:
report.status = "PASS"
report.status_extended = (

View File

@@ -10,6 +10,7 @@ class dynamodb_accelerator_cluster_encryption_enabled(Check):
report.resource_id = cluster.name
report.resource_arn = cluster.arn
report.region = cluster.region
report.resource_tags = cluster.tags
report.status = "FAIL"
report.status_extended = f"DynamoDB cluster {cluster.name} does not have encryption at rest enabled."
if cluster.encryption:

View File

@@ -1,5 +1,7 @@
import threading
from dataclasses import dataclass
from typing import Optional
from pydantic import BaseModel
from prowler.lib.logger import logger
from prowler.lib.scan_filters.scan_filters import is_resource_filtered
@@ -18,6 +20,7 @@ class DynamoDB:
self.__threading_call__(self.__list_tables__)
self.__describe_table__()
self.__describe_continuous_backups__()
self.__list_tags_for_resource__()
def __get_session__(self):
return self.session
@@ -94,6 +97,20 @@ class DynamoDB:
f"{error.__class__.__name__}:{error.__traceback__.tb_lineno} -- {error}"
)
def __list_tags_for_resource__(self):
logger.info("DynamoDB - List Tags...")
try:
for table in self.tables:
regional_client = self.regional_clients[table.region]
response = regional_client.list_tags_of_resource(ResourceArn=table.arn)[
"Tags"
]
table.tags = response
except Exception as error:
logger.error(
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
################## DynamoDB DAX
class DAX:
@@ -105,6 +122,7 @@ class DAX:
self.regional_clients = generate_regional_clients(self.service, audit_info)
self.clusters = []
self.__threading_call__(self.__describe_clusters__)
self.__list_tags_for_resource__()
def __get_session__(self):
return self.session
@@ -137,10 +155,10 @@ class DAX:
encryption = True
self.clusters.append(
Cluster(
cluster["ClusterArn"],
cluster["ClusterName"],
encryption,
regional_client.region,
arn=cluster["ClusterArn"],
name=cluster["ClusterName"],
encryption=encryption,
region=regional_client.region,
)
)
except Exception as error:
@@ -148,47 +166,32 @@ class DAX:
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
def __list_tags_for_resource__(self):
logger.info("DAX - List Tags...")
try:
for cluster in self.clusters:
regional_client = self.regional_clients[cluster.region]
response = regional_client.list_tags(ResourceName=cluster.name)["Tags"]
cluster.tags = response
except Exception as error:
logger.error(
f"{regional_client.region} -- {error.__class__.__name__}[{error.__traceback__.tb_lineno}]: {error}"
)
@dataclass
class Table:
class Table(BaseModel):
arn: str
name: str
encryption_type: str
kms_arn: str
pitr: bool
encryption_type: Optional[str]
kms_arn: Optional[str]
pitr: bool = False
region: str
def __init__(
self,
arn,
name,
encryption_type,
kms_arn,
region,
):
self.arn = arn
self.name = name
self.encryption_type = encryption_type
self.kms_arn = kms_arn
self.pitr = False
self.region = region
tags: Optional[list] = []
@dataclass
class Cluster:
class Cluster(BaseModel):
arn: str
name: str
encryption: str
encryption: bool
region: str
def __init__(
self,
arn,
name,
encryption,
region,
):
self.arn = arn
self.name = name
self.encryption = encryption
self.region = region
tags: Optional[list] = []

View File

@@ -9,6 +9,7 @@ class dynamodb_tables_kms_cmk_encryption_enabled(Check):
report = Check_Report_AWS(self.metadata())
report.resource_id = table.name
report.resource_arn = table.arn
report.resource_tags = table.tags
report.region = table.region
report.status = "FAIL"
report.status_extended = (

View File

@@ -9,6 +9,7 @@ class dynamodb_tables_pitr_enabled(Check):
report = Check_Report_AWS(self.metadata())
report.resource_id = table.name
report.resource_arn = table.arn
report.resource_tags = table.tags
report.region = table.region
report.status = "FAIL"
report.status_extended = f"DynamoDB table {table.name} does not have point-in-time recovery enabled."

View File

@@ -210,3 +210,7 @@ class Test_CloudFormation_Service:
assert cloudformation.stacks[0].is_nested_stack is False
assert cloudformation.stacks[0].root_nested_stack == ""
assert cloudformation.stacks[0].region == AWS_REGION
assert cloudformation.stacks[0].tags == [
{"Key": "Tag1", "Value": "Value1"},
{"Key": "Tag2", "Value": "Value2"},
]

View File

@@ -31,7 +31,7 @@ class Test_cloudformation_stacks_termination_protection_enabled:
Stack(
arn="arn:aws:cloudformation:eu-west-1:123456789012:stack/Test-Stack/796c8d26-b390-41d7-a23c-0702c4e78b60",
name=stack_name,
outputs="",
outputs=[],
region=AWS_REGION,
)
]
@@ -68,7 +68,7 @@ class Test_cloudformation_stacks_termination_protection_enabled:
Stack(
arn="arn:aws:cloudformation:eu-west-1:123456789012:stack/Test-Stack/796c8d26-b390-41d7-a23c-0702c4e78b60",
name=stack_name,
outputs="",
outputs=[],
region=AWS_REGION,
)
]

View File

@@ -133,6 +133,14 @@ def mock_make_api_call(self, operation_name, kwarg):
},
"ETag": "",
}
if operation_name == "ListTagsForResource":
return {
"Tags": {
"Items": [
{"Key": "test", "Value": "test"},
]
}
}
return make_api_call(self, operation_name, kwarg)
@@ -247,3 +255,7 @@ class Test_CloudFront_Service:
].default_cache_config.field_level_encryption_id
== "enabled"
)
assert cloudfront.distributions[cloudfront_distribution_id].tags == [
{"Key": "test", "Value": "test"},
]

View File

@@ -63,7 +63,6 @@ class Test_Cloudtrail_Service:
@mock_cloudtrail
@mock_s3
def test_describe_trails(self):
cloudtrail_client_us_east_1 = client("cloudtrail", region_name="us-east-1")
s3_client_us_east_1 = client("s3", region_name="us-east-1")
cloudtrail_client_eu_west_1 = client("cloudtrail", region_name="eu-west-1")
@@ -78,10 +77,20 @@ class Test_Cloudtrail_Service:
CreateBucketConfiguration={"LocationConstraint": "eu-west-1"},
)
cloudtrail_client_us_east_1.create_trail(
Name=trail_name_us, S3BucketName=bucket_name_us, IsMultiRegionTrail=False
Name=trail_name_us,
S3BucketName=bucket_name_us,
IsMultiRegionTrail=False,
TagsList=[
{"Key": "test", "Value": "test"},
],
)
cloudtrail_client_eu_west_1.create_trail(
Name=trail_name_eu, S3BucketName=bucket_name_eu, IsMultiRegionTrail=False
Name=trail_name_eu,
S3BucketName=bucket_name_eu,
IsMultiRegionTrail=False,
TagsList=[
{"Key": "test", "Value": "test"},
],
)
audit_info = self.set_mocked_audit_info()
cloudtrail = Cloudtrail(audit_info)
@@ -101,6 +110,9 @@ class Test_Cloudtrail_Service:
trail.s3_bucket == bucket_name_eu
or trail.s3_bucket == bucket_name_us
)
assert trail.tags == [
{"Key": "test", "Value": "test"},
]
@mock_cloudtrail
@mock_s3

View File

@@ -72,6 +72,13 @@ def mock_make_api_call(self, operation_name, kwarg):
],
}
if operation_name == "ListTagsForResource":
return {
"tags": [
{"key": "test", "value": "test"},
]
}
return make_api_call(self, operation_name, kwarg)
@@ -116,6 +123,9 @@ class Test_CodeArtifact_Service:
assert codeartifact.repositories
assert codeartifact.repositories["test-repository"]
assert codeartifact.repositories["test-repository"].name == "test-repository"
assert codeartifact.repositories["test-repository"].tags == [
{"key": "test", "value": "test"},
]
assert (
codeartifact.repositories["test-repository"].arn
== f"arn:aws:codebuild:{AWS_REGION}:{DEFAULT_ACCOUNT_ID}:repository/test-repository"

View File

@@ -13,6 +13,7 @@ class Test_config_recorder_all_regions_enabled:
from prowler.providers.aws.services.config.config_service import Config
current_audit_info.audited_partition = "aws"
current_audit_info.audited_account = "012345678912"
current_audit_info.audited_regions = ["eu-west-1", "us-east-1"]
with mock.patch(
@@ -44,7 +45,7 @@ class Test_config_recorder_all_regions_enabled:
from prowler.providers.aws.services.config.config_service import Config
current_audit_info.audited_partition = "aws"
current_audit_info.audited_regions = ["eu-west-1", "us-east-1"]
current_audit_info.audited_regions = [AWS_REGION]
with mock.patch(
"prowler.providers.aws.services.config.config_recorder_all_regions_enabled.config_recorder_all_regions_enabled.config_client",
@@ -57,7 +58,7 @@ class Test_config_recorder_all_regions_enabled:
check = config_recorder_all_regions_enabled()
result = check.execute()
assert len(result) == 2
assert len(result) == 1
# Search for the recorder just created
for recorder in result:
if recorder.resource_id:
@@ -85,7 +86,7 @@ class Test_config_recorder_all_regions_enabled:
from prowler.providers.aws.services.config.config_service import Config
current_audit_info.audited_partition = "aws"
current_audit_info.audited_regions = ["eu-west-1", "us-east-1"]
current_audit_info.audited_regions = [AWS_REGION]
with mock.patch(
"prowler.providers.aws.services.config.config_recorder_all_regions_enabled.config_recorder_all_regions_enabled.config_client",
@@ -98,7 +99,7 @@ class Test_config_recorder_all_regions_enabled:
check = config_recorder_all_regions_enabled()
result = check.execute()
assert len(result) == 2
assert len(result) == 1
# Search for the recorder just created
for recorder in result:
if recorder.resource_id:

View File

@@ -92,7 +92,12 @@ def mock_make_api_call(self, operation_name, kwarg):
"ManualSnapshotsLimitReached": True,
}
}
if operation_name == "ListTagsForResource":
return {
"Tags": [
{"Key": "string", "Value": "string"},
],
}
return make_api_call(self, operation_name, kwarg)
@@ -145,6 +150,9 @@ class Test_DirectoryService_Service:
)
assert directoryservice.directories["d-12345a1b2"].name == "test-directory"
assert directoryservice.directories["d-12345a1b2"].region == AWS_REGION
assert directoryservice.directories["d-12345a1b2"].tags == [
{"Key": "string", "Value": "string"},
]
assert (
directoryservice.directories[
"d-12345a1b2"

View File

@@ -121,6 +121,9 @@ class Test_DynamoDB_Service:
{"AttributeName": "app", "KeyType": "RANGE"},
],
BillingMode="PAY_PER_REQUEST",
Tags=[
{"Key": "test", "Value": "test"},
],
)["TableDescription"]
# DynamoDB client for this test class
audit_info = self.set_mocked_audit_info()
@@ -129,6 +132,9 @@ class Test_DynamoDB_Service:
assert dynamo.tables[0].arn == table["TableArn"]
assert dynamo.tables[0].name == "test1"
assert dynamo.tables[0].region == AWS_REGION
assert dynamo.tables[0].tags == [
{"Key": "test", "Value": "test"},
]
# Test DynamoDB Describe Table
@mock_dynamodb
@@ -174,6 +180,9 @@ class Test_DynamoDB_Service:
ReplicationFactor=3,
IamRoleArn=iam_role_arn,
SSESpecification={"Enabled": True},
Tags=[
{"Key": "test", "Value": "test"},
],
)
dax_client.create_cluster(
ClusterName="daxcluster2",
@@ -181,6 +190,9 @@ class Test_DynamoDB_Service:
ReplicationFactor=3,
IamRoleArn=iam_role_arn,
SSESpecification={"Enabled": True},
Tags=[
{"Key": "test", "Value": "test"},
],
)
# DAX client for this test class
audit_info = self.set_mocked_audit_info()
@@ -190,3 +202,9 @@ class Test_DynamoDB_Service:
assert dax.clusters[1].name == "daxcluster2"
assert dax.clusters[0].region == AWS_REGION
assert dax.clusters[1].region == AWS_REGION
assert dax.clusters[0].tags == [
{"Key": "test", "Value": "test"},
]
assert dax.clusters[1].tags == [
{"Key": "test", "Value": "test"},
]