# -*- coding: utf-8 -*-

"""
oss2.xml_utils
~~~~~~~~~~~~~~

XML处理相关。

主要包括两类接口：
    - parse_开头的函数：用来解析服务器端返回的XML
    - to_开头的函数：用来生成发往服务器端的XML

"""
import logging
import xml.etree.ElementTree as ElementTree
import json

from .models import (SimplifiedObjectInfo,
                     SimplifiedBucketInfo,
                     PartInfo,
                     MultipartUploadInfo,
                     LifecycleRule,
                     LifecycleExpiration,
                     CorsRule,
                     LiveChannelInfoTarget,
                     LiveChannelInfo,
                     LiveRecord,
                     LiveChannelVideoStat,
                     LiveChannelAudioStat,
                     Owner,
                     AccessControlList,
                     AbortMultipartUpload,
                     StorageTransition,
                     Tagging,
                     TaggingRule,
                     ServerSideEncryptionRule,
                     ListObjectVersionsResult,
                     ObjectVersionInfo,
                     DeleteMarkerInfo,
                     BatchDeleteObjectVersionResult,
                     BucketWebsite,
                     RoutingRule,
                     Condition,
                     ConditionInlcudeHeader,
                     Redirect,
                     RedirectMirrorHeaders,
                     MirrorHeadersSet,
                     REDIRECT_TYPE_MIRROR,
                     REDIRECT_TYPE_EXTERNAL,
                     REDIRECT_TYPE_INTERNAL,
                     REDIRECT_TYPE_ALICDN,
                     NoncurrentVersionStorageTransition,
                     NoncurrentVersionExpiration,
                     AsyncFetchTaskConfiguration,
                     InventoryConfiguration,
                     InventoryFilter,
                     InventorySchedule,
                     InventoryDestination,
                     InventoryBucketDestination,
                     InventoryServerSideEncryptionKMS,
                     InventoryServerSideEncryptionOSS,
                     LocationTransferType,
                     BucketReplicationProgress,
                     ReplicationRule,
                     CnameInfo,
                     CertificateInfo,
                     ReplicationRule,
                     MetaQueryFile,
                     AggregationsInfo,
                     OSSTaggingInfo,
                     OSSUserMetaInfo,
                     AggregationGroupInfo,
                     AccessMonitorInfo,
                     LifecycleFilter,
                     FilterNot,
                     FilterNotTag,
                     BucketStyleInfo,
                     RegionInfo,
                     CallbackPolicyInfo,
                     DataRedundancyTransitionInfo,
                     ListAccessPointResult,
                     AccessPointVpcConfiguration,
                     AccessPointEndpoints,
                     AccessPointInfo,
                     PublicAccessBlockConfiguration,
                     ResourcePoolInfo,
                     ResourcePoolBucketInfo,
                     RequesterQoSInfo,
                     QoSConfiguration)

from .select_params import (SelectJsonTypes, SelectParameters)

from .compat import urlunquote, to_unicode, to_string
from .utils import iso8601_to_unixtime, date_to_iso8601, iso8601_to_date
from . import utils
import base64
from .exceptions import SelectOperationClientError

logger = logging.getLogger(__name__)

def _find_tag(parent, path):
    child = parent.find(path)
    if child is None:
        raise RuntimeError("parse xml: " + path + " could not be found under " + parent.tag)

    if child.text is None:
        return ''

    return to_string(child.text)

def _find_tag_with_default(parent, path, default_value):
    child = parent.find(path)
    if child is None:
        return default_value

    if child.text is None:
        return ''

    return to_string(child.text)

def _find_bool(parent, path):
    text = _find_tag(parent, path)
    if text == 'true':
        return True
    elif text == 'false':
        return False
    else:
        raise RuntimeError("parse xml: value of " + path + " is not a boolean under " + parent.tag)


def _find_int(parent, path):
    return int(_find_tag(parent, path))


def _find_object(parent, path, url_encoded):
    name = _find_tag(parent, path)
    if url_encoded:
        return urlunquote(name)
    else:
        return name


def _find_all_tags(parent, tag):
    return [to_string(node.text) or '' for node in parent.findall(tag)]


def _is_url_encoding(root):
    node = root.find('EncodingType')
    if node is not None and to_string(node.text) == 'url':
        return True
    else:
        return False


def _node_to_string(root):
    return ElementTree.tostring(root, encoding='utf-8')


def _add_node_list(parent, tag, entries):
    for e in entries:
        _add_text_child(parent, tag, e)


def _add_text_child(parent, tag, text):
    ElementTree.SubElement(parent, tag).text = to_unicode(text)

def _add_node_child(parent, tag):
    return ElementTree.SubElement(parent, tag)

def parse_list_objects(result, body):
    root = ElementTree.fromstring(body)
    url_encoded = _is_url_encoding(root)
    result.is_truncated = _find_bool(root, 'IsTruncated')
    if result.is_truncated:
        result.next_marker = _find_object(root, 'NextMarker', url_encoded)

    for contents_node in root.findall('Contents'):
        owner = None
        if contents_node.find("Owner") is not None:
            owner = Owner(_find_tag(contents_node, 'Owner/DisplayName'), _find_tag(contents_node, 'Owner/ID'))
        result.object_list.append(SimplifiedObjectInfo(
            _find_object(contents_node, 'Key', url_encoded),
            iso8601_to_unixtime(_find_tag(contents_node, 'LastModified')),
            _find_tag(contents_node, 'ETag').strip('"'),
            _find_tag(contents_node, 'Type'),
            int(_find_tag(contents_node, 'Size')),
            _find_tag(contents_node, 'StorageClass'),
            owner,
            _find_tag_with_default(contents_node, 'RestoreInfo', None)
        ))

    for prefix_node in root.findall('CommonPrefixes'):
        result.prefix_list.append(_find_object(prefix_node, 'Prefix', url_encoded))

    return result


def parse_list_objects_v2(result, body):
    root = ElementTree.fromstring(body)
    url_encoded = _is_url_encoding(root)
    result.is_truncated = _find_bool(root, 'IsTruncated')
    if result.is_truncated:
        result.next_continuation_token = _find_object(root, 'NextContinuationToken', url_encoded)

    for contents_node in root.findall('Contents'):
        owner = None
        if contents_node.find("Owner") is not None:
            owner = Owner(_find_tag(contents_node, 'Owner/DisplayName'), _find_tag(contents_node, 'Owner/ID'))
        result.object_list.append(SimplifiedObjectInfo(
            _find_object(contents_node, 'Key', url_encoded),
            iso8601_to_unixtime(_find_tag(contents_node, 'LastModified')),
            _find_tag(contents_node, 'ETag').strip('"'),
            _find_tag(contents_node, 'Type'),
            int(_find_tag(contents_node, 'Size')),
            _find_tag(contents_node, 'StorageClass'),
            owner,
            _find_tag_with_default(contents_node, 'RestoreInfo', None)
        ))

    for prefix_node in root.findall('CommonPrefixes'):
        result.prefix_list.append(_find_object(prefix_node, 'Prefix', url_encoded))

    return result


def parse_list_buckets(result, body):
    root = ElementTree.fromstring(body)

    if root.find('IsTruncated') is None:
        result.is_truncated = False
    else:
        result.is_truncated = _find_bool(root, 'IsTruncated')

    if result.is_truncated:
        result.next_marker = _find_tag(root, 'NextMarker')

    if root.find('Owner') is not None:
        result.owner = Owner(_find_tag_with_default(root, 'Owner/DisplayName', None), _find_tag_with_default(root, 'Owner/ID', None))

    for bucket_node in root.findall('Buckets/Bucket'):
        result.buckets.append(SimplifiedBucketInfo(
            _find_tag(bucket_node, 'Name'),
            _find_tag(bucket_node, 'Location'),
            iso8601_to_unixtime(_find_tag(bucket_node, 'CreationDate')),
            _find_tag(bucket_node, 'ExtranetEndpoint'),
            _find_tag(bucket_node, 'IntranetEndpoint'),
            _find_tag(bucket_node, 'StorageClass'),
            _find_tag_with_default(bucket_node, 'Region', None),
            _find_tag_with_default(bucket_node, 'ResourceGroupId', None),
        ))

    return result


def parse_init_multipart_upload(result, body):
    root = ElementTree.fromstring(body)
    result.upload_id = _find_tag(root, 'UploadId')

    return result


def parse_list_multipart_uploads(result, body):
    root = ElementTree.fromstring(body)

    url_encoded = _is_url_encoding(root)

    result.is_truncated = _find_bool(root, 'IsTruncated')
    result.next_key_marker = _find_object(root, 'NextKeyMarker', url_encoded)
    result.next_upload_id_marker = _find_tag(root, 'NextUploadIdMarker')

    for upload_node in root.findall('Upload'):
        result.upload_list.append(MultipartUploadInfo(
            _find_object(upload_node, 'Key', url_encoded),
            _find_tag(upload_node, 'UploadId'),
            iso8601_to_unixtime(_find_tag(upload_node, 'Initiated'))
        ))

    for prefix_node in root.findall('CommonPrefixes'):
        result.prefix_list.append(_find_object(prefix_node, 'Prefix', url_encoded))

    return result


def parse_list_parts(result, body):
    root = ElementTree.fromstring(body)

    result.is_truncated = _find_bool(root, 'IsTruncated')
    result.next_marker = _find_tag(root, 'NextPartNumberMarker')
    for part_node in root.findall('Part'):
        result.parts.append(PartInfo(
            _find_int(part_node, 'PartNumber'),
            _find_tag(part_node, 'ETag').strip('"'),
            size=_find_int(part_node, 'Size'),
            last_modified=iso8601_to_unixtime(_find_tag(part_node, 'LastModified'))
        ))

    return result


def parse_batch_delete_objects(result, body):
    if not body:
        return result 
    root = ElementTree.fromstring(body)
    url_encoded = _is_url_encoding(root)

    for deleted_node in root.findall('Deleted'):
        key = _find_object(deleted_node, 'Key', url_encoded)

        result.deleted_keys.append(key)

        versionid_node = deleted_node.find('VersionId')
        versionid = None
        if versionid_node is not None:
            versionid = _find_tag(deleted_node, 'VersionId')

        delete_marker_node = deleted_node.find('DeleteMarker')
        delete_marker = False
        if delete_marker_node is not None:
            delete_marker = _find_bool(deleted_node, 'DeleteMarker')

        marker_versionid_node = deleted_node.find('DeleteMarkerVersionId')
        delete_marker_versionid = ''
        if marker_versionid_node is not None:
            delete_marker_versionid = _find_tag(deleted_node, 'DeleteMarkerVersionId')
        result.delete_versions.append(BatchDeleteObjectVersionResult(key, versionid, delete_marker, delete_marker_versionid))

    return result


def parse_get_bucket_acl(result, body):
    root = ElementTree.fromstring(body)
    result.acl = _find_tag(root, 'AccessControlList/Grant')

    return result


def parse_get_object_acl(result, body):
    root = ElementTree.fromstring(body)
    result.acl = _find_tag(root, 'AccessControlList/Grant')

    return result


def parse_get_bucket_location(result, body):
    result.location = to_string(ElementTree.fromstring(body).text)
    return result


def parse_get_bucket_logging(result, body):
    root = ElementTree.fromstring(body)

    if root.find('LoggingEnabled/TargetBucket') is not None:
        result.target_bucket = _find_tag(root, 'LoggingEnabled/TargetBucket')

    if root.find('LoggingEnabled/TargetPrefix') is not None:
        result.target_prefix = _find_tag(root, 'LoggingEnabled/TargetPrefix')

    return result


def parse_get_bucket_stat(result, body):
    root = ElementTree.fromstring(body)

    result.storage_size_in_bytes = _find_int(root, 'Storage')
    result.object_count = _find_int(root, 'ObjectCount')
    result.multi_part_upload_count = int(_find_tag_with_default(root, 'MultipartUploadCount', 0))
    if root.find('LiveChannelCount') is not None:
        result.live_channel_count = int(_find_tag(root, 'LiveChannelCount'))
    if root.find('LastModifiedTime') is not None:
        result.last_modified_time = int(_find_tag(root, 'LastModifiedTime'))
    if root.find('StandardStorage') is not None:
        result.standard_storage = int(_find_tag(root, 'StandardStorage'))
    if root.find('StandardObjectCount') is not None:
        result.standard_object_count = int(_find_tag(root, 'StandardObjectCount'))
    if root.find('InfrequentAccessStorage') is not None:
        result.infrequent_access_storage = int(_find_tag(root, 'InfrequentAccessStorage'))
    if root.find('InfrequentAccessRealStorage') is not None:
        result.infrequent_access_real_storage = int(_find_tag(root, 'InfrequentAccessRealStorage'))
    if root.find('InfrequentAccessObjectCount') is not None:
        result.infrequent_access_object_count = int(_find_tag(root, 'InfrequentAccessObjectCount'))
    if root.find('ArchiveStorage') is not None:
        result.archive_storage = int(_find_tag(root, 'ArchiveStorage'))
    if root.find('ArchiveRealStorage') is not None:
        result.archive_real_storage = int(_find_tag(root, 'ArchiveRealStorage'))
    if root.find('ArchiveObjectCount') is not None:
        result.archive_object_count = int(_find_tag(root, 'ArchiveObjectCount'))
    if root.find('ColdArchiveStorage') is not None:
        result.cold_archive_storage = int(_find_tag(root, 'ColdArchiveStorage'))
    if root.find('ColdArchiveRealStorage') is not None:
        result.cold_archive_real_storage = int(_find_tag(root, 'ColdArchiveRealStorage'))
    if root.find('ColdArchiveObjectCount') is not None:
        result.cold_archive_object_count = int(_find_tag(root, 'ColdArchiveObjectCount'))
    if root.find('MultipartPartCount') is not None:
        result.multipart_part_count = int(_find_tag(root, 'MultipartPartCount'))
    if root.find('DeleteMarkerCount') is not None:
        result.delete_marker_count = int(_find_tag(root, 'DeleteMarkerCount'))
    if root.find('DeepColdArchiveStorage') is not None:
        result.deep_cold_archive_storage = int(_find_tag(root, 'DeepColdArchiveStorage'))
    if root.find('DeepColdArchiveRealStorage') is not None:
        result.deep_cold_archive_real_storage = int(_find_tag(root, 'DeepColdArchiveRealStorage'))
    if root.find('DeepColdArchiveObjectCount') is not None:
        result.deep_cold_archive_object_count = int(_find_tag(root, 'DeepColdArchiveObjectCount'))

    return result


def parse_get_bucket_info(result, body):
    root = ElementTree.fromstring(body)

    result.name = _find_tag(root, 'Bucket/Name')
    result.creation_date = _find_tag(root, 'Bucket/CreationDate')
    result.storage_class = _find_tag(root, 'Bucket/StorageClass')
    result.extranet_endpoint = _find_tag(root, 'Bucket/ExtranetEndpoint')
    result.intranet_endpoint = _find_tag(root, 'Bucket/IntranetEndpoint')
    result.location = _find_tag(root, 'Bucket/Location')
    result.owner = Owner(_find_tag(root, 'Bucket/Owner/DisplayName'), _find_tag(root, 'Bucket/Owner/ID'))
    result.acl = AccessControlList(_find_tag(root, 'Bucket/AccessControlList/Grant'))
    result.comment = _find_tag_with_default(root, 'Bucket/Comment', None)
    result.versioning_status = _find_tag_with_default(root, 'Bucket/Versioning', None)
    result.data_redundancy_type = _find_tag_with_default(root, 'Bucket/DataRedundancyType', None)
    result.access_monitor = _find_tag_with_default(root, 'Bucket/AccessMonitor', None)
    result.transfer_acceleration = _find_tag_with_default(root, 'Bucket/TransferAcceleration', None)
    result.cross_region_replication = _find_tag_with_default(root, 'Bucket/CrossRegionReplication', None)
    result.resource_group_id = _find_tag_with_default(root, 'Bucket/ResourceGroupId', None)

    server_side_encryption = root.find("Bucket/ServerSideEncryptionRule")
    if server_side_encryption is None:
        result.bucket_encryption_rule = None
    else:
        result.bucket_encryption_rule = _parse_bucket_encryption_info(server_side_encryption)

    return result

def _parse_bucket_encryption_info(node):

    rule = ServerSideEncryptionRule()

    rule.sse_algorithm = _find_tag(node,"SSEAlgorithm")
    
    if rule.sse_algorithm == "None":
        rule.kms_master_keyid = None
        rule.sse_algorithm = None
        return rule

    kmsnode = node.find("KMSMasterKeyID")
    if kmsnode is None or kmsnode.text is None:
        rule.kms_master_keyid = None 
    else:
        rule.kms_master_keyid = to_string(kmsnode.text)

    kms_data_encryption_node = node.find("KMSDataEncryption")
    if kms_data_encryption_node is None or kms_data_encryption_node.text is None:
        rule.kms_data_encryption = None
    else:
        rule.kms_data_encryption = to_string(kms_data_encryption_node.text)

    return rule

def parse_get_bucket_referer(result, body):
    root = ElementTree.fromstring(body)

    result.allow_empty_referer = _find_bool(root, 'AllowEmptyReferer')
    result.referers = _find_all_tags(root, 'RefererList/Referer')

    if root.find("AllowTruncateQueryString") is not None:
        result.allow_truncate_query_string = _find_bool(root, 'AllowTruncateQueryString')
    if root.find("RefererBlacklist/Referer") is not None:
        result.black_referers = _find_all_tags(root, 'RefererBlacklist/Referer')

    return result

def parse_condition_include_header(include_header_node):
    key = _find_tag(include_header_node, 'Key')
    equals = _find_tag(include_header_node, 'Equals')
    include_header = ConditionInlcudeHeader(key, equals)

    return include_header

def parse_routing_rule_condition(condition_node):
    if condition_node.find('KeyPrefixEquals') is not None:
        key_prefix_equals = _find_tag(condition_node, 'KeyPrefixEquals')
    if condition_node.find('HttpErrorCodeReturnedEquals') is not None:
        http_err_code_return_equals = _find_int(condition_node, 'HttpErrorCodeReturnedEquals');
        
    include_header_list = []
    if condition_node.find('IncludeHeader') is not None:
        for include_header_node in condition_node.findall('IncludeHeader'):
            include_header = parse_condition_include_header(include_header_node)
            include_header_list.append(include_header)

    condition = Condition(key_prefix_equals, http_err_code_return_equals, include_header_list)

    return condition

def parse_mirror_headers(mirror_headers_node):
    if mirror_headers_node is None:
        return None
    
    pass_all = None
    if mirror_headers_node.find('PassAll') is not None:
        pass_all = _find_bool(mirror_headers_node, 'PassAll')
    
    pass_list = _find_all_tags(mirror_headers_node, 'Pass')
    remove_list = _find_all_tags(mirror_headers_node, 'Remove')
    set_list = []
    for set_node in mirror_headers_node.findall('Set'):
        key = _find_tag(set_node, 'Key')
        value = _find_tag(set_node, 'Value')
        mirror_headers_set = MirrorHeadersSet(key, value)
        set_list.append(mirror_headers_set)

    redirect_mirror_headers = RedirectMirrorHeaders(pass_all, pass_list, remove_list, set_list)

    return redirect_mirror_headers

def parse_routing_rule_redirect(redirect_node):
    redirect_type = None
    pass_query_string = None
    replace_key_with = None
    replace_key_prefix_with = None
    proto = None
    host_name = None
    http_redirect_code = None
    mirror_url = None
    mirror_url_slave = None
    mirror_url_probe = None
    mirror_pass_query_string = None
    mirror_check_md5 = None
    mirror_follow_redirect = None
    mirror_headers = None

    # common args
    redirect_type = _find_tag(redirect_node, 'RedirectType')

    if redirect_node.find('PassQueryString') is not None:
        pass_query_string = _find_bool(redirect_node, 'PassQueryString')

    # External, AliCDN
    if redirect_type in [REDIRECT_TYPE_EXTERNAL, REDIRECT_TYPE_ALICDN]:
        if redirect_node.find('Protocol') is not None:
            proto = _find_tag(redirect_node, 'Protocol')

        if redirect_node.find('HostName') is not None:
            host_name = _find_tag(redirect_node, 'HostName')

        if redirect_node.find('HttpRedirectCode') is not None:
            http_redirect_code = _find_int(redirect_node, 'HttpRedirectCode')

    # External, AliCDN, Internal
    if redirect_type in [REDIRECT_TYPE_EXTERNAL, REDIRECT_TYPE_ALICDN, REDIRECT_TYPE_INTERNAL]:
        if redirect_node.find('ReplaceKeyWith') is not None:
            replace_key_with = _find_tag(redirect_node, 'ReplaceKeyWith')
        if redirect_node.find('ReplaceKeyPrefixWith') is not None:
            replace_key_prefix_with = _find_tag(redirect_node, 'ReplaceKeyPrefixWith')

    # Mirror
    elif redirect_type == REDIRECT_TYPE_MIRROR:
        if redirect_node.find('MirrorURL') is not None:
            mirror_url = _find_tag(redirect_node, 'MirrorURL')

        if redirect_node.find('MirrorURLSlave') is not None:
            mirror_url_slave = _find_tag(redirect_node, 'MirrorURLSlave')

        if redirect_node.find('MirrorURLProbe') is not None:
            mirror_url_probe = _find_tag(redirect_node, 'MirrorURLProbe')

        if redirect_node.find('MirrorPassQueryString') is not None:
            mirror_pass_query_string = _find_bool(redirect_node, 'MirrorPassQueryString')

        if redirect_node.find('MirrorCheckMd5') is not None:
            mirror_check_md5 = _find_bool(redirect_node, 'MirrorCheckMd5')

        if redirect_node.find('MirrorFollowRedirect') is not None:
            mirror_follow_redirect = _find_bool(redirect_node, 'MirrorFollowRedirect')

        mirror_headers = parse_mirror_headers(redirect_node.find('MirrorHeaders'))

    redirect = Redirect(redirect_type=redirect_type, proto=proto, host_name=host_name, replace_key_with=replace_key_with, 
                    replace_key_prefix_with=replace_key_prefix_with, http_redirect_code=http_redirect_code, 
                    pass_query_string=pass_query_string, mirror_url=mirror_url,mirror_url_slave=mirror_url_slave, 
                    mirror_url_probe=mirror_url_probe, mirror_pass_query_string=mirror_pass_query_string, 
                    mirror_follow_redirect=mirror_follow_redirect, mirror_check_md5=mirror_check_md5, 
                    mirror_headers=mirror_headers)

    return redirect

def parse_get_bucket_website(result, body):
    root = ElementTree.fromstring(body)
    result.index_file = _find_tag_with_default(root, 'IndexDocument/Suffix', None)
    result.error_file = _find_tag_with_default(root, 'ErrorDocument/Key', None)

    if root.find('RoutingRules') is None:
        return result

    routing_rules_node = root.find('RoutingRules')

    for rule_node in routing_rules_node.findall('RoutingRule'):
        rule_num = _find_int(rule_node, 'RuleNumber')
        condition = parse_routing_rule_condition(rule_node.find('Condition'))
        redirect = parse_routing_rule_redirect(rule_node.find('Redirect'))
        rule = RoutingRule(rule_num, condition, redirect);
        result.rules.append(rule)

    return result


def parse_create_live_channel(result, body):
    root = ElementTree.fromstring(body)

    result.play_url = _find_tag(root, 'PlayUrls/Url')
    result.publish_url = _find_tag(root, 'PublishUrls/Url')

    return result


def parse_get_live_channel(result, body):
    root = ElementTree.fromstring(body)

    result.status = _find_tag(root, 'Status')
    result.description = _find_tag(root, 'Description')

    target = LiveChannelInfoTarget()
    target.type = _find_tag(root, 'Target/Type')
    target.frag_duration = _find_tag(root, 'Target/FragDuration')
    target.frag_count = _find_tag(root, 'Target/FragCount')
    target.playlist_name = _find_tag(root, 'Target/PlaylistName')

    result.target = target

    return result


def parse_list_live_channel(result, body):
    root = ElementTree.fromstring(body)

    result.prefix = _find_tag(root, 'Prefix')
    result.marker = _find_tag(root, 'Marker')
    result.max_keys = _find_int(root, 'MaxKeys')
    result.is_truncated = _find_bool(root, 'IsTruncated')
    
    if result.is_truncated:
        result.next_marker = _find_tag(root, 'NextMarker')

    channels = root.findall('LiveChannel')
    for channel in channels:
        tmp = LiveChannelInfo()
        tmp.name = _find_tag(channel, 'Name')
        tmp.description = _find_tag(channel, 'Description')
        tmp.status = _find_tag(channel, 'Status')
        tmp.last_modified = iso8601_to_unixtime(_find_tag(channel, 'LastModified'))
        tmp.play_url = _find_tag(channel, 'PlayUrls/Url')
        tmp.publish_url = _find_tag(channel, 'PublishUrls/Url')

        result.channels.append(tmp)

    return result


def parse_stat_video(video_node, video):
    video.width = _find_int(video_node, 'Width')
    video.height = _find_int(video_node, 'Height')
    video.frame_rate = _find_int(video_node, 'FrameRate')
    video.bandwidth = _find_int(video_node, 'Bandwidth')
    video.codec = _find_tag(video_node, 'Codec')


def parse_stat_audio(audio_node, audio):
    audio.bandwidth = _find_int(audio_node, 'Bandwidth')
    audio.sample_rate = _find_int(audio_node, 'SampleRate')
    audio.codec = _find_tag(audio_node, 'Codec')


def parse_live_channel_stat(result, body):
    root = ElementTree.fromstring(body)

    result.status = _find_tag(root, 'Status')
    if root.find('RemoteAddr') is not None:
        result.remote_addr = _find_tag(root, 'RemoteAddr')
    if root.find('ConnectedTime') is not None:
        result.connected_time = iso8601_to_unixtime(_find_tag(root, 'ConnectedTime'))

    video_node = root.find('Video')
    audio_node = root.find('Audio')

    if video_node is not None:
        result.video = LiveChannelVideoStat()
        parse_stat_video(video_node, result.video)
    if audio_node is not None:
        result.audio = LiveChannelAudioStat()
        parse_stat_audio(audio_node, result.audio)

    return result


def parse_live_channel_history(result, body):
    root = ElementTree.fromstring(body)

    records = root.findall('LiveRecord')
    for record in records:
        tmp = LiveRecord()
        tmp.start_time = iso8601_to_unixtime(_find_tag(record, 'StartTime'))
        tmp.end_time = iso8601_to_unixtime(_find_tag(record, 'EndTime'))
        tmp.remote_addr = _find_tag(record, 'RemoteAddr')
        result.records.append(tmp)

    return result


def parse_lifecycle_expiration(expiration_node):
    if expiration_node is None:
        return None

    expiration = LifecycleExpiration()

    if expiration_node.find('Days') is not None:
        expiration.days = _find_int(expiration_node, 'Days')
    elif expiration_node.find('Date') is not None:
        expiration.date = iso8601_to_date(_find_tag(expiration_node, 'Date'))
    elif expiration_node.find('CreatedBeforeDate') is not None:
        expiration.created_before_date = iso8601_to_date(_find_tag(expiration_node, 'CreatedBeforeDate'))
    elif expiration_node.find('ExpiredObjectDeleteMarker') is not None:
        expiration.expired_detete_marker = _find_bool(expiration_node, 'ExpiredObjectDeleteMarker')

    return expiration


def parse_lifecycle_abort_multipart_upload(abort_multipart_upload_node):
    if abort_multipart_upload_node is None:
        return None
    abort_multipart_upload = AbortMultipartUpload()

    if abort_multipart_upload_node.find('Days') is not None:
        abort_multipart_upload.days = _find_int(abort_multipart_upload_node, 'Days')
    elif abort_multipart_upload_node.find('CreatedBeforeDate') is not None:
        abort_multipart_upload.created_before_date = iso8601_to_date(_find_tag(abort_multipart_upload_node,
                                                                               'CreatedBeforeDate'))
    return abort_multipart_upload


def parse_lifecycle_storage_transitions(storage_transition_nodes):
    storage_transitions = []
    for storage_transition_node in storage_transition_nodes:
        storage_class = _find_tag(storage_transition_node, 'StorageClass')
        storage_transition = StorageTransition(storage_class=storage_class)
        if storage_transition_node.find('Days') is not None:
            storage_transition.days = _find_int(storage_transition_node, 'Days')
        elif storage_transition_node.find('CreatedBeforeDate') is not None:
            storage_transition.created_before_date = iso8601_to_date(_find_tag(storage_transition_node,
                                                                               'CreatedBeforeDate'))
        if storage_transition_node.find('IsAccessTime') is not None:
            storage_transition.is_access_time = _find_bool(storage_transition_node, 'IsAccessTime')
        if storage_transition_node.find('ReturnToStdWhenVisit') is not None:
            storage_transition.return_to_std_when_visit = _find_bool(storage_transition_node, 'ReturnToStdWhenVisit')
        if storage_transition_node.find('AllowSmallFile') is not None:
            storage_transition.allow_small_file = _find_bool(storage_transition_node, 'AllowSmallFile')

        storage_transitions.append(storage_transition)

    return storage_transitions

def parse_lifecycle_object_taggings(lifecycle_tagging_nodes):
    
    if lifecycle_tagging_nodes is None or \
        len(lifecycle_tagging_nodes) == 0: 
        return None 
    
    tagging_rule = TaggingRule()
    for tag_node in lifecycle_tagging_nodes:
        key = _find_tag(tag_node, 'Key')
        value = _find_tag(tag_node, 'Value')
        tagging_rule.add(key, value)

    return Tagging(tagging_rule)

def parse_lifecycle_version_expiration(version_expiration_node):
    if version_expiration_node is None:
        return None

    noncurrent_days = _find_int(version_expiration_node, 'NoncurrentDays')
    expiration = NoncurrentVersionExpiration(noncurrent_days)

    return expiration

def parse_lifecycle_verison_storage_transitions(version_storage_transition_nodes):
    version_storage_transitions = []
    for transition_node in version_storage_transition_nodes:
        storage_class = _find_tag(transition_node, 'StorageClass')
        non_crurrent_days = _find_int(transition_node, 'NoncurrentDays')
        version_storage_transition = NoncurrentVersionStorageTransition(non_crurrent_days, storage_class)
        if transition_node.find('IsAccessTime') is not None:
            version_storage_transition.is_access_time = _find_bool(transition_node, 'IsAccessTime')
        if transition_node.find('ReturnToStdWhenVisit') is not None:
            version_storage_transition.return_to_std_when_visit = _find_bool(transition_node, 'ReturnToStdWhenVisit')
        if transition_node.find('AllowSmallFile') is not None:
            version_storage_transition.allow_small_file = _find_bool(transition_node, 'AllowSmallFile')

        version_storage_transitions.append(version_storage_transition)

    return version_storage_transitions

def parse_get_bucket_lifecycle(result, body):

    root = ElementTree.fromstring(body)
    url_encoded = _is_url_encoding(root)

    for rule_node in root.findall('Rule'):
        expiration = parse_lifecycle_expiration(rule_node.find('Expiration'))
        abort_multipart_upload = parse_lifecycle_abort_multipart_upload(rule_node.find('AbortMultipartUpload'))
        storage_transitions = parse_lifecycle_storage_transitions(rule_node.findall('Transition'))
        tagging = parse_lifecycle_object_taggings(rule_node.findall('Tag'))
        noncurrent_version_expiration = parse_lifecycle_version_expiration(rule_node.find('NoncurrentVersionExpiration'))
        noncurrent_version_sotrage_transitions = parse_lifecycle_verison_storage_transitions(rule_node.findall('NoncurrentVersionTransition'))
        lifecycle_filter = parse_lifecycle_filter(rule_node.find('Filter'))

        rule = LifecycleRule(
            _find_tag(rule_node, 'ID'),
            _find_tag(rule_node, 'Prefix'),
            status=_find_tag(rule_node, 'Status'),
            expiration=expiration,
            abort_multipart_upload=abort_multipart_upload,
            storage_transitions=storage_transitions,
            tagging=tagging,
            noncurrent_version_expiration = noncurrent_version_expiration,
            noncurrent_version_sotrage_transitions = noncurrent_version_sotrage_transitions,
            atime_base=int(_find_tag_with_default(rule_node, 'AtimeBase', 0)),
            filter=lifecycle_filter
            )
        result.rules.append(rule)

    return result


def parse_get_bucket_cors(result, body):
    root = ElementTree.fromstring(body)

    for rule_node in root.findall('CORSRule'):
        rule = CorsRule()
        rule.allowed_origins = _find_all_tags(rule_node, 'AllowedOrigin')
        rule.allowed_methods = _find_all_tags(rule_node, 'AllowedMethod')
        rule.allowed_headers = _find_all_tags(rule_node, 'AllowedHeader')
        rule.expose_headers = _find_all_tags(rule_node, 'ExposeHeader')

        max_age_node = rule_node.find('MaxAgeSeconds')
        if max_age_node is not None:
            rule.max_age_seconds = int(max_age_node.text)

        result.rules.append(rule)

    if root.find('ResponseVary') is not None:
        result.response_vary = _find_bool(root, 'ResponseVary')

    return result


def to_complete_upload_request(parts):
    root = ElementTree.Element('CompleteMultipartUpload')
    for p in parts:
        part_node = ElementTree.SubElement(root, "Part")
        _add_text_child(part_node, 'PartNumber', str(p.part_number))
        _add_text_child(part_node, 'ETag', '"{0}"'.format(p.etag))

    return _node_to_string(root)


def to_batch_delete_objects_request(keys, quiet):
    root_node = ElementTree.Element('Delete')

    _add_text_child(root_node, 'Quiet', str(quiet).lower())

    for key in keys:
        object_node = ElementTree.SubElement(root_node, 'Object')
        _add_text_child(object_node, 'Key', key)

    return _node_to_string(root_node)

def to_batch_delete_objects_version_request(objectVersions, quiet):

    root_node = ElementTree.Element('Delete')

    _add_text_child(root_node, 'Quiet', str(quiet).lower())

    objectVersionList = objectVersions.object_version_list

    for ver in objectVersionList:
        object_node = ElementTree.SubElement(root_node, 'Object')
        _add_text_child(object_node, 'Key', ver.key)
        if ver.versionid != '':
            _add_text_child(object_node, 'VersionId', ver.versionid)

    return _node_to_string(root_node)


def to_put_bucket_config(bucket_config):
    root = ElementTree.Element('CreateBucketConfiguration')

    _add_text_child(root, 'StorageClass', bucket_config.storage_class)

    if bucket_config.data_redundancy_type is not None:
        _add_text_child(root, 'DataRedundancyType', bucket_config.data_redundancy_type)

    return _node_to_string(root)


def to_put_bucket_logging(bucket_logging):
    root = ElementTree.Element('BucketLoggingStatus')

    if bucket_logging.target_bucket:
        logging_node = ElementTree.SubElement(root, 'LoggingEnabled')
        _add_text_child(logging_node, 'TargetBucket', bucket_logging.target_bucket)
        _add_text_child(logging_node, 'TargetPrefix', bucket_logging.target_prefix)

    return _node_to_string(root)


def to_put_bucket_referer(bucket_referer):
    root = ElementTree.Element('RefererConfiguration')

    _add_text_child(root, 'AllowEmptyReferer', str(bucket_referer.allow_empty_referer).lower())
    list_node = ElementTree.SubElement(root, 'RefererList')

    for r in bucket_referer.referers:
        _add_text_child(list_node, 'Referer', r)

    if bucket_referer.allow_truncate_query_string is not None:
        _add_text_child(root, 'AllowTruncateQueryString', str(bucket_referer.allow_truncate_query_string).lower())

    if bucket_referer.black_referers:
        black_referer_node = ElementTree.SubElement(root, 'RefererBlacklist')

        for r in bucket_referer.black_referers:
            _add_text_child(black_referer_node, 'Referer', r)

    return _node_to_string(root)


def to_put_bucket_website(bucket_website):
    root = ElementTree.Element('WebsiteConfiguration')

    index_node = ElementTree.SubElement(root, 'IndexDocument')
    _add_text_child(index_node, 'Suffix', bucket_website.index_file)

    error_node = ElementTree.SubElement(root, 'ErrorDocument')
    _add_text_child(error_node, 'Key', bucket_website.error_file)

    if len(bucket_website.rules) == 0:
        return _node_to_string(root)

    rules_node = ElementTree.SubElement(root, "RoutingRules")

    for rule in bucket_website.rules:
        rule_node = ElementTree.SubElement(rules_node, 'RoutingRule')
        _add_text_child(rule_node, 'RuleNumber', str(rule.rule_num))

        condition_node = ElementTree.SubElement(rule_node, 'Condition')
        
        if rule.condition.key_prefix_equals is not None:
            _add_text_child(condition_node, 'KeyPrefixEquals', rule.condition.key_prefix_equals)
        if rule.condition.http_err_code_return_equals is not None:    
            _add_text_child(condition_node, 'HttpErrorCodeReturnedEquals', 
                str(rule.condition.http_err_code_return_equals))
       
        for header in rule.condition.include_header_list:
            include_header_node = ElementTree.SubElement(condition_node, 'IncludeHeader')
            _add_text_child(include_header_node, 'Key', header.key)
            _add_text_child(include_header_node, 'Equals', header.equals)

        if rule.redirect is not None:    
            redirect_node = ElementTree.SubElement(rule_node, 'Redirect')

            # common
            _add_text_child(redirect_node, 'RedirectType', rule.redirect.redirect_type)
            
            if rule.redirect.pass_query_string is not None:
                _add_text_child(redirect_node, 'PassQueryString', str(rule.redirect.pass_query_string))          

            # External, AliCDN
            if rule.redirect.redirect_type in [REDIRECT_TYPE_EXTERNAL, REDIRECT_TYPE_ALICDN]:
                if rule.redirect.proto is not None:
                    _add_text_child(redirect_node, 'Protocol', rule.redirect.proto)
                if rule.redirect.host_name is not None:
                    _add_text_child(redirect_node, 'HostName', rule.redirect.host_name)
                if rule.redirect.http_redirect_code is not None:
                    _add_text_child(redirect_node, 'HttpRedirectCode', str(rule.redirect.http_redirect_code))

            # External, AliCDN, Internal
            if rule.redirect.redirect_type in [REDIRECT_TYPE_EXTERNAL, REDIRECT_TYPE_ALICDN, REDIRECT_TYPE_INTERNAL]:
                if rule.redirect.replace_key_with is not None:
                    _add_text_child(redirect_node, 'ReplaceKeyWith', rule.redirect.replace_key_with)
                if rule.redirect.replace_key_prefix_with is not None:
                    _add_text_child(redirect_node, 'ReplaceKeyPrefixWith', rule.redirect.replace_key_prefix_with)  

            # Mirror
            elif rule.redirect.redirect_type == REDIRECT_TYPE_MIRROR: 
                if rule.redirect.mirror_url is not None:
                    _add_text_child(redirect_node, 'MirrorURL', rule.redirect.mirror_url)
                if rule.redirect.mirror_url_slave is not None:
                    _add_text_child(redirect_node, 'MirrorURLSlave', rule.redirect.mirror_url_slave)
                if rule.redirect.mirror_url_probe is not None:
                    _add_text_child(redirect_node, 'MirrorURLProbe', rule.redirect.mirror_url_probe)
                if rule.redirect.mirror_pass_query_string is not None:
                    _add_text_child(redirect_node, 'MirrorPassQueryString', str(rule.redirect.mirror_pass_query_string))
                if rule.redirect.mirror_follow_redirect is not None:
                    _add_text_child(redirect_node, 'MirrorFollowRedirect', str(rule.redirect.mirror_follow_redirect))
                if rule.redirect.mirror_check_md5 is not None:
                    _add_text_child(redirect_node, 'MirrorCheckMd5', str(rule.redirect.mirror_check_md5))

                if rule.redirect.mirror_headers is not None:
                    mirror_headers_node = ElementTree.SubElement(redirect_node, 'MirrorHeaders')

                    if rule.redirect.mirror_headers.pass_all is not None:
                        _add_text_child(mirror_headers_node, 'PassAll', str(rule.redirect.mirror_headers.pass_all))

                    for pass_param in rule.redirect.mirror_headers.pass_list:
                        _add_text_child(mirror_headers_node, 'Pass', pass_param)   
                    for remove_param in rule.redirect.mirror_headers.remove_list:
                        _add_text_child(mirror_headers_node, 'Remove', remove_param)
                    for set_param in rule.redirect.mirror_headers.set_list:
                        set_node = ElementTree.SubElement(mirror_headers_node, 'Set')
                        _add_text_child(set_node, 'Key', set_param.key)
                        _add_text_child(set_node, 'Value', set_param.value)

    return _node_to_string(root)


def to_put_bucket_lifecycle(bucket_lifecycle):
    root = ElementTree.Element('LifecycleConfiguration')

    for rule in bucket_lifecycle.rules:
        rule_node = ElementTree.SubElement(root, 'Rule')
        _add_text_child(rule_node, 'ID', rule.id)
        _add_text_child(rule_node, 'Prefix', rule.prefix)
        _add_text_child(rule_node, 'Status', rule.status)

        expiration = rule.expiration
        if expiration:
            expiration_node = ElementTree.SubElement(rule_node, 'Expiration')

            if expiration.days is not None:
                _add_text_child(expiration_node, 'Days', str(expiration.days))
            elif expiration.date is not None:
                _add_text_child(expiration_node, 'Date', date_to_iso8601(expiration.date))
            elif expiration.created_before_date is not None:
                _add_text_child(expiration_node, 'CreatedBeforeDate', date_to_iso8601(expiration.created_before_date))
            elif expiration.expired_detete_marker is not None:
                _add_text_child(expiration_node, 'ExpiredObjectDeleteMarker', str(expiration.expired_detete_marker))

        abort_multipart_upload = rule.abort_multipart_upload
        if abort_multipart_upload:
            abort_multipart_upload_node = ElementTree.SubElement(rule_node, 'AbortMultipartUpload')
            if abort_multipart_upload.days is not None:
                _add_text_child(abort_multipart_upload_node, 'Days', str(abort_multipart_upload.days))
            elif abort_multipart_upload.created_before_date is not None:
                _add_text_child(abort_multipart_upload_node, 'CreatedBeforeDate',
                                date_to_iso8601(abort_multipart_upload.created_before_date))

        storage_transitions = rule.storage_transitions
        if storage_transitions:
            for storage_transition in storage_transitions:
                storage_transition_node = ElementTree.SubElement(rule_node, 'Transition')
                _add_text_child(storage_transition_node, 'StorageClass', str(storage_transition.storage_class))
                if storage_transition.is_access_time is not None:
                    _add_text_child(storage_transition_node, 'IsAccessTime', str(storage_transition.is_access_time).lower())
                if storage_transition.return_to_std_when_visit is not None:
                    _add_text_child(storage_transition_node, 'ReturnToStdWhenVisit', str(storage_transition.return_to_std_when_visit).lower())
                if storage_transition.allow_small_file is not None:
                    _add_text_child(storage_transition_node, 'AllowSmallFile', str(storage_transition.allow_small_file).lower())
                if storage_transition.days is not None:
                    _add_text_child(storage_transition_node, 'Days', str(storage_transition.days))
                elif storage_transition.created_before_date is not None:
                    _add_text_child(storage_transition_node, 'CreatedBeforeDate',
                                    date_to_iso8601(storage_transition.created_before_date))

        tagging = rule.tagging
        if tagging:
            tagging_rule = tagging.tag_set.tagging_rule
            for key in tagging.tag_set.tagging_rule:
                tag_node = ElementTree.SubElement(rule_node, 'Tag')
                _add_text_child(tag_node, 'Key', key)
                _add_text_child(tag_node, 'Value', tagging_rule[key])

        noncurrent_version_expiration = rule.noncurrent_version_expiration
        if noncurrent_version_expiration is not None:
            version_expiration_node = ElementTree.SubElement(rule_node, 'NoncurrentVersionExpiration')
            _add_text_child(version_expiration_node, 'NoncurrentDays', str(noncurrent_version_expiration.noncurrent_days))

        noncurrent_version_sotrage_transitions = rule.noncurrent_version_sotrage_transitions
        if noncurrent_version_sotrage_transitions is not None:
            for noncurrent_version_sotrage_transition in noncurrent_version_sotrage_transitions:
                version_transition_node = ElementTree.SubElement(rule_node, 'NoncurrentVersionTransition')
                _add_text_child(version_transition_node, 'NoncurrentDays', str(noncurrent_version_sotrage_transition.noncurrent_days))
                _add_text_child(version_transition_node, 'StorageClass', str(noncurrent_version_sotrage_transition.storage_class))
                if noncurrent_version_sotrage_transition.is_access_time is not None:
                    _add_text_child(version_transition_node, 'IsAccessTime', str(noncurrent_version_sotrage_transition.is_access_time).lower())
                if noncurrent_version_sotrage_transition.return_to_std_when_visit is not None:
                    _add_text_child(version_transition_node, 'ReturnToStdWhenVisit', str(noncurrent_version_sotrage_transition.return_to_std_when_visit).lower())
                if noncurrent_version_sotrage_transition.allow_small_file is not None:
                    _add_text_child(version_transition_node, 'AllowSmallFile', str(noncurrent_version_sotrage_transition.allow_small_file).lower())

        if rule.filter:
            filter_node = ElementTree.SubElement(rule_node, "Filter")
            if rule.filter.object_size_greater_than:
                _add_text_child(filter_node, 'ObjectSizeGreaterThan', str(rule.filter.object_size_greater_than))
            if rule.filter.object_size_less_than:
                _add_text_child(filter_node, 'ObjectSizeLessThan', str(rule.filter.object_size_less_than))
            if rule.filter.filter_not:
                for not_arg in rule.filter.filter_not:
                    not_node = ElementTree.SubElement(filter_node, 'Not')

                    _add_text_child(not_node, 'Prefix', not_arg.prefix)
                    if not_arg.tag:
                        tag_node = ElementTree.SubElement(not_node, 'Tag')
                        _add_text_child(tag_node, 'Key', not_arg.tag.key)
                        _add_text_child(tag_node, 'Value', not_arg.tag.value)

    return _node_to_string(root)


def to_put_bucket_cors(bucket_cors):
    root = ElementTree.Element('CORSConfiguration')

    for rule in bucket_cors.rules:
        rule_node = ElementTree.SubElement(root, 'CORSRule')
        _add_node_list(rule_node, 'AllowedOrigin', rule.allowed_origins)
        _add_node_list(rule_node, 'AllowedMethod', rule.allowed_methods)
        _add_node_list(rule_node, 'AllowedHeader', rule.allowed_headers)
        _add_node_list(rule_node, 'ExposeHeader', rule.expose_headers)

        if rule.max_age_seconds is not None:
            _add_text_child(rule_node, 'MaxAgeSeconds', str(rule.max_age_seconds))

    if bucket_cors.response_vary is not None:
        _add_text_child(root, 'ResponseVary', str(bucket_cors.response_vary).lower())

    return _node_to_string(root)

def to_create_live_channel(live_channel):
    root = ElementTree.Element('LiveChannelConfiguration')

    _add_text_child(root, 'Description', live_channel.description)
    _add_text_child(root, 'Status', live_channel.status)
    target_node = _add_node_child(root, 'Target')

    _add_text_child(target_node, 'Type', live_channel.target.type)
    _add_text_child(target_node, 'FragDuration', str(live_channel.target.frag_duration))
    _add_text_child(target_node, 'FragCount', str(live_channel.target.frag_count))
    _add_text_child(target_node, 'PlaylistName', str(live_channel.target.playlist_name))

    return _node_to_string(root)

def to_select_object(sql, select_params):
    if (select_params is not None and 'Json_Type' in select_params):
        return to_select_json_object(sql, select_params)
    else:
        return to_select_csv_object(sql, select_params)

def to_select_csv_object(sql, select_params):
    root = ElementTree.Element('SelectRequest')
    _add_text_child(root, 'Expression', base64.b64encode(str.encode(sql)))
    input_ser = ElementTree.SubElement(root, 'InputSerialization')
    output_ser = ElementTree.SubElement(root, 'OutputSerialization')
    csv = ElementTree.SubElement(input_ser, 'CSV')
    out_csv = ElementTree.SubElement(output_ser, 'CSV')
    options = ElementTree.SubElement(root, 'Options')
   
    if (select_params is None):
        return _node_to_string(root)
    
    for key, value in select_params.items():
        if SelectParameters.CsvHeaderInfo == key:
            _add_text_child(csv, 'FileHeaderInfo', value)
        elif SelectParameters.CommentCharacter == key:
            _add_text_child(csv, SelectParameters.CommentCharacter, base64.b64encode(str.encode(value)))
        elif SelectParameters.RecordDelimiter == key:
            _add_text_child(csv, SelectParameters.RecordDelimiter, base64.b64encode(str.encode(value)))
        elif SelectParameters.OutputRecordDelimiter == key:
            _add_text_child(out_csv, SelectParameters.RecordDelimiter, base64.b64encode(str.encode(value)))
        elif SelectParameters.FieldDelimiter == key:
            _add_text_child(csv, SelectParameters.FieldDelimiter, base64.b64encode(str.encode(value)))
        elif SelectParameters.OutputFieldDelimiter == key:
            _add_text_child(out_csv, SelectParameters.FieldDelimiter, base64.b64encode(str.encode(value)))
        elif SelectParameters.QuoteCharacter == key:
            _add_text_child(csv, SelectParameters.QuoteCharacter, base64.b64encode(str.encode(value)))
        elif SelectParameters.SplitRange == key:
            _add_text_child(csv, 'Range', utils._make_split_range_string(value))
        elif SelectParameters.LineRange == key:
            _add_text_child(csv, 'Range', utils._make_line_range_string(value))
        elif SelectParameters.CompressionType == key:
            _add_text_child(input_ser, SelectParameters.CompressionType, str(value))
        elif SelectParameters.KeepAllColumns == key:
            _add_text_child(output_ser, SelectParameters.KeepAllColumns, str(value))
        elif SelectParameters.OutputRawData == key:
            _add_text_child(output_ser, SelectParameters.OutputRawData, str(value))
        elif SelectParameters.EnablePayloadCrc == key:
            _add_text_child(output_ser, SelectParameters.EnablePayloadCrc, str(value))
        elif SelectParameters.OutputHeader == key:
            _add_text_child(output_ser, SelectParameters.OutputHeader, str(value))
        elif SelectParameters.SkipPartialDataRecord == key:
            _add_text_child(options, SelectParameters.SkipPartialDataRecord, str(value))
        elif SelectParameters.MaxSkippedRecordsAllowed == key:
            _add_text_child(options, SelectParameters.MaxSkippedRecordsAllowed, str(value))
        elif SelectParameters.AllowQuotedRecordDelimiter == key:
            _add_text_child(csv, SelectParameters.AllowQuotedRecordDelimiter, str(value))
        else:
            raise SelectOperationClientError("The select_params contains unsupported key " + key, "")

    return _node_to_string(root)

def to_select_json_object(sql, select_params):
    root = ElementTree.Element('SelectRequest')
    _add_text_child(root, 'Expression', base64.b64encode(str.encode(sql)))
    input_ser = ElementTree.SubElement(root, 'InputSerialization')
    output_ser = ElementTree.SubElement(root, 'OutputSerialization')
    json = ElementTree.SubElement(input_ser, 'JSON')
    out_json = ElementTree.SubElement(output_ser, 'JSON')
    options = ElementTree.SubElement(root, 'Options')
    is_doc = select_params[SelectParameters.Json_Type] == SelectJsonTypes.DOCUMENT
    _add_text_child(json, 'Type', select_params[SelectParameters.Json_Type])

    for key, value in select_params.items(): 
        if SelectParameters.SplitRange == key and is_doc == False:
            _add_text_child(json, 'Range', utils._make_split_range_string(value))
        elif SelectParameters.LineRange == key and is_doc == False:
            _add_text_child(json, 'Range', utils._make_line_range_string(value))
        elif SelectParameters.CompressionType == key:
            _add_text_child(input_ser, SelectParameters.CompressionType, value)
        elif SelectParameters.OutputRawData == key:
            _add_text_child(output_ser, SelectParameters.OutputRawData, str(value))
        elif SelectParameters.EnablePayloadCrc == key:
            _add_text_child(output_ser, SelectParameters.EnablePayloadCrc, str(value))
        elif SelectParameters.OutputRecordDelimiter == key:
            _add_text_child(out_json, SelectParameters.RecordDelimiter, base64.b64encode(str.encode(value)))
        elif SelectParameters.SkipPartialDataRecord == key:
            _add_text_child(options, SelectParameters.SkipPartialDataRecord, str(value))
        elif SelectParameters.MaxSkippedRecordsAllowed == key:
            _add_text_child(options, SelectParameters.MaxSkippedRecordsAllowed, str(value))
        elif SelectParameters.ParseJsonNumberAsString == key:
            _add_text_child(json, SelectParameters.ParseJsonNumberAsString, str(value))
        else:
            if key != SelectParameters.Json_Type:
                raise SelectOperationClientError("The select_params contains unsupported key " + key, "")

    return _node_to_string(root)

def to_get_select_object_meta(meta_param):
    if meta_param is not None and SelectParameters.Json_Type in meta_param:
        if meta_param[SelectParameters.Json_Type] != SelectJsonTypes.LINES:
            raise SelectOperationClientError("Json_Type can only be 'LINES' for creating meta", "")
        else:
            return to_get_select_json_object_meta(meta_param)
    else:
        return to_get_select_csv_object_meta(meta_param)

def to_get_select_csv_object_meta(csv_meta_param):
    root = ElementTree.Element('CsvMetaRequest')
    input_ser = ElementTree.SubElement(root, 'InputSerialization')
    csv = ElementTree.SubElement(input_ser, 'CSV')
    if (csv_meta_param is None):
        return _node_to_string(root)
    
    for key, value in csv_meta_param.items():
        if SelectParameters.RecordDelimiter == key:
            _add_text_child(csv, SelectParameters.RecordDelimiter, base64.b64encode(str.encode(value)))
        elif SelectParameters.FieldDelimiter == key:
            _add_text_child(csv, SelectParameters.FieldDelimiter, base64.b64encode(str.encode(value)))
        elif SelectParameters.QuoteCharacter == key:
            _add_text_child(csv, SelectParameters.QuoteCharacter, base64.b64encode(str.encode(value)))
        elif SelectParameters.CompressionType == key:
            _add_text_child(input_ser, SelectParameters.CompressionType, base64.b64encode(str.encode(value)))
        elif SelectParameters.OverwriteIfExists == key:
            _add_text_child(root, SelectParameters.OverwriteIfExists, str(value))
        else:
           raise SelectOperationClientError("The csv_meta_param contains unsupported key " + key, "") 

    return _node_to_string(root)

def to_get_select_json_object_meta(json_meta_param):
    root = ElementTree.Element('JsonMetaRequest')
    input_ser = ElementTree.SubElement(root, 'InputSerialization')
    json = ElementTree.SubElement(input_ser, 'JSON')
    _add_text_child(json, 'Type', json_meta_param[SelectParameters.Json_Type]) # Json_Type是必须的
  
    for key, value in json_meta_param.items():
        if SelectParameters.OverwriteIfExists == key:
            _add_text_child(root, SelectParameters.OverwriteIfExists, str(value))
        elif SelectParameters.CompressionType == key:
             _add_text_child(input_ser, SelectParameters.CompressionType, base64.b64encode(str.encode(value)))
        else:
            if SelectParameters.Json_Type != key:
                raise SelectOperationClientError("The json_meta_param contains unsupported key " + key, "")
            
    return _node_to_string(root)

def to_put_tagging(object_tagging):
    root = ElementTree.Element("Tagging")
    tag_set = ElementTree.SubElement(root, "TagSet")

    for item in object_tagging.tag_set.tagging_rule:
        tag_xml = ElementTree.SubElement(tag_set, "Tag")
        _add_text_child(tag_xml, 'Key', item)
        _add_text_child(tag_xml, 'Value', object_tagging.tag_set.tagging_rule[item])

    return _node_to_string(root)

def parse_get_tagging(result, body):
    root = ElementTree.fromstring(body)
    url_encoded = _is_url_encoding(root)
    tagset_node = root.find('TagSet')

    if tagset_node is None:
        return result

    tagging_rules = TaggingRule()
    for tag_node in tagset_node.findall('Tag'):
        key = _find_object(tag_node, 'Key', url_encoded)
        value = _find_object(tag_node, 'Value', url_encoded)
        tagging_rules.add(key, value)
    
    result.tag_set = tagging_rules
    return result

def to_put_bucket_encryption(rule):
    root = ElementTree.Element("ServerSideEncryptionRule")
    apply_node = ElementTree.SubElement(root, "ApplyServerSideEncryptionByDefault")

    _add_text_child(apply_node, "SSEAlgorithm", rule.sse_algorithm)

    if rule.kms_master_keyid:
        _add_text_child(apply_node, "KMSMasterKeyID", rule.kms_master_keyid)

    if rule.kms_data_encryption:
        _add_text_child(apply_node, "KMSDataEncryption", rule.kms_data_encryption)

    return _node_to_string(root)

def parse_get_bucket_encryption(result, body):
    root = ElementTree.fromstring(body)
    apply_node = root.find('ApplyServerSideEncryptionByDefault')

    result.sse_algorithm = _find_tag(apply_node, "SSEAlgorithm")

    kmsnode = apply_node.find('KMSMasterKeyID')
    if kmsnode is None or kmsnode.text is None:
        result.kms_master_keyid = None 
    else:
        result.kms_master_keyid = to_string(kmsnode.text)

    kms_data_encryption_node = apply_node.find('KMSDataEncryption')
    if kms_data_encryption_node is None or kms_data_encryption_node.text is None:
        result.kms_data_encryption = None
    else:
        result.kms_data_encryption = to_string(kms_data_encryption_node.text)

    return result

def parse_list_object_versions(result, body):
    root = ElementTree.fromstring(body)
    url_encoded = _is_url_encoding(root)
    result.is_truncated = _find_bool(root, 'IsTruncated')
    if result.is_truncated:
        result.next_key_marker = _find_object(root, 'NextKeyMarker', url_encoded)
        result.next_versionid_marker = _find_object(root, "NextVersionIdMarker", url_encoded)

    result.name = _find_tag(root, "Name")
    result.prefix = _find_object(root, "Prefix", url_encoded)
    result.key_marker = _find_object(root, "KeyMarker", url_encoded)
    result.versionid_marker = _find_object(root, "VersionIdMarker", url_encoded)
    result.max_keys = _find_int(root, "MaxKeys")
    result.delimiter = _find_object(root, "Delimiter", url_encoded)

    for delete_marker in root.findall("DeleteMarker"):
        deleteInfo = DeleteMarkerInfo()
        deleteInfo.key = _find_object(delete_marker, "Key", url_encoded)
        deleteInfo.versionid = _find_tag(delete_marker, "VersionId")
        deleteInfo.is_latest = _find_bool(delete_marker, "IsLatest")
        deleteInfo.last_modified = iso8601_to_unixtime(_find_tag(delete_marker, "LastModified"))
        deleteInfo.owner.id = _find_tag(delete_marker, "Owner/ID")
        deleteInfo.owner.display_name = _find_tag(delete_marker, "Owner/DisplayName")
        result.delete_marker.append(deleteInfo)

    for version in root.findall("Version"):
        versionInfo = ObjectVersionInfo()
        versionInfo.key = _find_object(version, "Key", url_encoded)
        versionInfo.versionid = _find_tag(version, "VersionId")
        versionInfo.is_latest = _find_bool(version, "IsLatest")
        versionInfo.last_modified = iso8601_to_unixtime(_find_tag(version, "LastModified"))
        versionInfo.owner.id = _find_tag(version, "Owner/ID")
        versionInfo.owner.display_name = _find_tag(version, "Owner/DisplayName")
        versionInfo.type = _find_tag(version, "Type")
        versionInfo.storage_class = _find_tag(version, "StorageClass")
        versionInfo.size = _find_int(version, "Size")
        versionInfo.etag = _find_tag(version, "ETag").strip('"')
        versionInfo.restore_info = _find_tag_with_default(version, 'RestoreInfo', None)

        result.versions.append(versionInfo)

    for common_prefix in root.findall("CommonPrefixes"):
        result.common_prefix.append(_find_object(common_prefix, "Prefix", url_encoded))

    return result

def to_put_bucket_versioning(bucket_version_config):
    root = ElementTree.Element('VersioningConfiguration')

    _add_text_child(root, 'Status', str(bucket_version_config.status))

    return _node_to_string(root)

def parse_get_bucket_versioning(result, body):
    root = ElementTree.fromstring(body)

    status_node = root.find("Status")
    if status_node is None:
        result.status = None
    else:
        result.status = _find_tag(root, "Status")

    return result

def to_put_bucket_request_payment(payer):
    root = ElementTree.Element('RequestPaymentConfiguration')

    _add_text_child(root, 'Payer', payer)

    return _node_to_string(root)

def parse_get_bucket_request_payment(result, body):
    root = ElementTree.fromstring(body)

    result.payer = _find_tag(root, 'Payer')
   
    return result

def to_put_qos_info(qos_info):
    root = ElementTree.Element("QoSConfiguration")

    if qos_info.total_upload_bw is not None:
        _add_text_child(root, "TotalUploadBandwidth", str(qos_info.total_upload_bw))
    if qos_info.intranet_upload_bw is not None:
        _add_text_child(root, "IntranetUploadBandwidth", str(qos_info.intranet_upload_bw))
    if qos_info.extranet_upload_bw is not None:
        _add_text_child(root, "ExtranetUploadBandwidth", str(qos_info.extranet_upload_bw))
    if qos_info.total_download_bw is not None:
        _add_text_child(root, "TotalDownloadBandwidth", str(qos_info.total_download_bw))
    if qos_info.intranet_download_bw is not None:
        _add_text_child(root, "IntranetDownloadBandwidth", str(qos_info.intranet_download_bw))
    if qos_info.extranet_download_bw is not None:
        _add_text_child(root, "ExtranetDownloadBandwidth", str(qos_info.extranet_download_bw))
    if qos_info.total_qps is not None:
        _add_text_child(root, "TotalQps", str(qos_info.total_qps))
    if qos_info.intranet_qps is not None:
        _add_text_child(root, "IntranetQps", str(qos_info.intranet_qps))
    if qos_info.extranet_qps is not None:
        _add_text_child(root, "ExtranetQps", str(qos_info.extranet_qps))

    return _node_to_string(root)

def parse_get_qos_info(result, body):
    """解析UserQosInfo 或者BucketQosInfo

    :UserQosInfo包含成员region,其他成员同BucketQosInfo
    """
    root = ElementTree.fromstring(body)

    if hasattr(result, 'region'):
        result.region = _find_tag(root, 'Region')

    result.total_upload_bw = _find_int(root, 'TotalUploadBandwidth')
    result.intranet_upload_bw = _find_int(root, 'IntranetUploadBandwidth')
    result.extranet_upload_bw = _find_int(root, 'ExtranetUploadBandwidth')
    result.total_download_bw = _find_int(root, 'TotalDownloadBandwidth')
    result.intranet_download_bw = _find_int(root, 'IntranetDownloadBandwidth')
    result.extranet_download_bw = _find_int(root, 'ExtranetDownloadBandwidth')
    result.total_qps = _find_int(root, 'TotalQps')
    result.intranet_qps = _find_int(root, 'IntranetQps')
    result.extranet_qps = _find_int(root, 'ExtranetQps')

    return result

def parse_get_bucket_user_qos(result, body):
    root = ElementTree.fromstring(body)

    result.storage_capacity = _find_int(root, 'StorageCapacity')

    return result

def to_put_bucket_user_qos(user_qos):
    root = ElementTree.Element('BucketUserQos')

    _add_text_child(root, 'StorageCapacity', str(user_qos.storage_capacity))

    return _node_to_string(root)


def to_put_async_fetch_task(task_config):
    root = ElementTree.Element('AsyncFetchTaskConfiguration')

    _add_text_child(root, 'Url', task_config.url)
    _add_text_child(root, 'Object', task_config.object_name)

    if task_config.host is not None:    
        _add_text_child(root, 'Host', task_config.host)
    if task_config.content_md5 is not None:
        _add_text_child(root, 'ContentMD5', task_config.content_md5)
    if task_config.callback is not None:
        _add_text_child(root, 'Callback', task_config.callback)
    if task_config.ignore_same_key is not None:
        _add_text_child(root, 'IgnoreSameKey', str(task_config.ignore_same_key).lower())
    if task_config.callback_when_failed is not None:
        _add_text_child(root, 'CallbackWhenFailed', str(task_config.callback_when_failed).lower())

    return _node_to_string(root)

def parse_put_async_fetch_task_result(result, body):
    root = ElementTree.fromstring(body)

    result.task_id = _find_tag(root, 'TaskId')

    return result

def _parse_async_fetch_task_configuration(task_info_node):
    url = _find_tag(task_info_node, 'Url')
    object_name = _find_tag(task_info_node, 'Object')
    host = _find_tag(task_info_node, 'Host')
    content_md5 = _find_tag(task_info_node, 'ContentMD5')
    callback = _find_tag(task_info_node, 'Callback')
    ignore_same_key = _find_bool(task_info_node, 'IgnoreSameKey')

    return AsyncFetchTaskConfiguration(url, object_name, host, content_md5, callback, ignore_same_key)

def parse_get_async_fetch_task_result(result, body):
    root = ElementTree.fromstring(body)

    result.task_id = _find_tag(root, 'TaskId')
    result.task_state = _find_tag(root, 'State')
    result.error_msg = _find_tag(root, 'ErrorMsg')
    result.task_config = _parse_async_fetch_task_configuration(root.find('TaskInfo'))

    return result

def to_put_inventory_configuration(inventory_config):
    root = ElementTree.Element("InventoryConfiguration")
    _add_text_child(root, "Id", inventory_config.inventory_id)

    if inventory_config.is_enabled is not None:
        _add_text_child(root, "IsEnabled", str(inventory_config.is_enabled))

    if inventory_config.included_object_versions is not None:
        _add_text_child(root, "IncludedObjectVersions", inventory_config.included_object_versions)

    if inventory_config.inventory_filter is not None:
        filter_node = ElementTree.SubElement(root, 'Filter')
        if inventory_config.inventory_filter.prefix is not None:
            _add_text_child(filter_node, "Prefix", inventory_config.inventory_filter.prefix)
        if inventory_config.inventory_filter.last_modify_begin_time_stamp is not None:
            _add_text_child(filter_node, "LastModifyBeginTimeStamp", str(inventory_config.inventory_filter.last_modify_begin_time_stamp))
        if inventory_config.inventory_filter.last_modify_end_time_stamp is not None:
            _add_text_child(filter_node, "LastModifyEndTimeStamp", str(inventory_config.inventory_filter.last_modify_end_time_stamp))
        if inventory_config.inventory_filter.lower_size_bound is not None:
            _add_text_child(filter_node, "LowerSizeBound", str(inventory_config.inventory_filter.lower_size_bound))
        if inventory_config.inventory_filter.upper_size_bound is not None:
            _add_text_child(filter_node, "UpperSizeBound", str(inventory_config.inventory_filter.upper_size_bound))
        if inventory_config.inventory_filter.storage_class is not None:
            _add_text_child(filter_node, "StorageClass", inventory_config.inventory_filter.storage_class)

    if inventory_config.inventory_schedule is not None and inventory_config.inventory_schedule.frequency is not None:
        schedule_node = ElementTree.SubElement(root, 'Schedule')
        _add_text_child(schedule_node, "Frequency", inventory_config.inventory_schedule.frequency)

    if inventory_config.optional_fields is not None:
        fields_node = ElementTree.SubElement(root, 'OptionalFields')
        for field in inventory_config.optional_fields:
            _add_text_child(fields_node, "Field", field)

    if inventory_config.inventory_destination is not None and inventory_config.inventory_destination.bucket_destination is not None:
        destin_node = ElementTree.SubElement(root, 'Destination')
        bucket_destin_node = ElementTree.SubElement(destin_node, 'OSSBucketDestination')
        bucket_destin = inventory_config.inventory_destination.bucket_destination

        if bucket_destin.account_id is not None:
            _add_text_child(bucket_destin_node, "AccountId", str(bucket_destin.account_id))

        if bucket_destin.role_arn is not None:
            _add_text_child(bucket_destin_node, "RoleArn", bucket_destin.role_arn)

        if bucket_destin.bucket is not None:
            _add_text_child(bucket_destin_node, "Bucket", "acs:oss:::" + bucket_destin.bucket)

        if bucket_destin.inventory_format is not None:
            _add_text_child(bucket_destin_node, "Format", bucket_destin.inventory_format)

        if bucket_destin.prefix is not None:
            _add_text_child(bucket_destin_node, "Prefix", bucket_destin.prefix)

        if bucket_destin.sse_kms_encryption is not None:
            encryption_node =  ElementTree.SubElement(bucket_destin_node, 'Encryption')
            sse_kms_node =  ElementTree.SubElement(encryption_node, 'SSE-KMS')
            _add_text_child(sse_kms_node, "KeyId", bucket_destin.sse_kms_encryption.key_id)
        elif bucket_destin.sse_oss_encryption is not None:
            encryption_node =  ElementTree.SubElement(bucket_destin_node, 'Encryption')
            _add_node_child(encryption_node, 'SSE-OSS')

    return _node_to_string(root)

def get_Inventory_configuration_from_element(elem):
    root = elem
    result = InventoryConfiguration()

    result.inventory_id = _find_tag(root, 'Id')
    result.is_enabled = _find_bool(root, 'IsEnabled')
    result.included_object_versions = _find_tag(root, 'IncludedObjectVersions')

    if root.find("Filter") is not None:
        prefix = None
        last_modify_begin_time_stamp = None
        last_modify_end_time_stamp = None
        lower_size_bound = None
        upper_size_bound = None
        storage_class = None
        if root.find("Filter/Prefix") is not None:
            prefix = _find_tag(root, 'Filter/Prefix')
        if root.find("Filter/LastModifyBeginTimeStamp") is not None:
            last_modify_begin_time_stamp = _find_tag_with_default(root, 'Filter/LastModifyBeginTimeStamp', None)
        if root.find("Filter/LastModifyEndTimeStamp") is not None:
            last_modify_end_time_stamp = _find_tag_with_default(root, 'Filter/LastModifyEndTimeStamp', None)
        if root.find("Filter/LowerSizeBound") is not None:
            lower_size_bound = _find_tag_with_default(root, 'Filter/LowerSizeBound', None)
        if root.find("Filter/UpperSizeBound") is not None:
            upper_size_bound = _find_tag_with_default(root, 'Filter/UpperSizeBound', None)
        if root.find("Filter/StorageClass") is not None:
            storage_class = _find_tag_with_default(root, 'Filter/StorageClass', None)
        result.inventory_filter = InventoryFilter(prefix=prefix, last_modify_begin_time_stamp=last_modify_begin_time_stamp,
                                                  last_modify_end_time_stamp=last_modify_end_time_stamp, lower_size_bound=lower_size_bound,
                                                  upper_size_bound=upper_size_bound, storage_class=storage_class)

    if root.find("Schedule/Frequency") is not None:
        result.inventory_schedule = InventorySchedule(_find_tag(root, 'Schedule/Frequency'))

    result.optional_fields =  _find_all_tags(root, "OptionalFields/Field")

    if root.find("Destination/OSSBucketDestination") is not None:
        bucket_distin_node = root.find("Destination/OSSBucketDestination")
        account_id = None
        role_arn = None
        bucket = None
        inventory_format = None
        prefix = None
        sse_kms_encryption = None
        sse_oss_encryption = None

        if bucket_distin_node.find('AccountId') is not None:
            account_id = _find_tag(bucket_distin_node, 'AccountId')
        if bucket_distin_node.find('RoleArn') is not None:
            role_arn = _find_tag(bucket_distin_node, 'RoleArn')
        if bucket_distin_node.find('Bucket') is not None:
            origin_bucket = _find_tag(bucket_distin_node, 'Bucket')
            if origin_bucket.startswith('acs:oss:::'):
                bucket = origin_bucket.replace('acs:oss:::', '')

        if bucket_distin_node.find('Format') is not None:
            inventory_format = _find_tag(bucket_distin_node, 'Format')
        if bucket_distin_node.find('Prefix') is not None:
            prefix = _find_tag(bucket_distin_node, 'Prefix')

        sse_kms_node = bucket_distin_node.find("Encryption/SSE-KMS")
        if sse_kms_node is not None:
            sse_kms_encryption = InventoryServerSideEncryptionKMS(_find_tag(sse_kms_node, 'KeyId'))
        elif bucket_distin_node.find("Encryption/SSE-OSS") is not None:
            sse_oss_encryption = InventoryServerSideEncryptionOSS()

        bucket_destination = InventoryBucketDestination(account_id=account_id, role_arn=role_arn, 
                bucket=bucket, inventory_format=inventory_format, prefix=prefix, 
                sse_kms_encryption=sse_kms_encryption, sse_oss_encryption=sse_oss_encryption)
 
        result.inventory_destination = InventoryDestination(bucket_destination)

    return result

def parse_get_bucket_inventory_configuration(result, body):
    root = ElementTree.fromstring(body)
    inventory_config = get_Inventory_configuration_from_element(root)

    result.inventory_id = inventory_config.inventory_id
    result.is_enabled = inventory_config.is_enabled
    result.included_object_versions = inventory_config.included_object_versions
    result.inventory_filter = inventory_config.inventory_filter
    result.inventory_schedule = inventory_config.inventory_schedule
    result.optional_fields = inventory_config.optional_fields
    result.inventory_destination = inventory_config.inventory_destination

    return result

def parse_list_bucket_inventory_configurations(result, body):
    root = ElementTree.fromstring(body)

    for inventory_config_node in root.findall("InventoryConfiguration"):
        inventory_config = get_Inventory_configuration_from_element(inventory_config_node)
        result.inventory_configurations.append(inventory_config)

    if root.find("ContinuationToken") is not None:
        result.continuaiton_token = _find_tag(root, "ContinuationToken")

    if root.find("IsTruncated") is not None:
        result.is_truncated = _find_bool(root, "IsTruncated")

    if root.find("NextContinuationToken") is not None:
        result.next_continuation_token = _find_tag(root, "NextContinuationToken")

    return result

def to_put_restore_config(restore_config):
    root = ElementTree.Element('RestoreRequest')

    _add_text_child(root, 'Days', str(restore_config.days))

    if restore_config.job_parameters is not None:
        job_parameters = restore_config.job_parameters
        job_parameters_node = ElementTree.SubElement(root, "JobParameters")
        if job_parameters.tier is not None:
            _add_text_child(job_parameters_node, 'Tier', job_parameters.tier)

    return _node_to_string(root)

def parse_get_bucket_worm_result(result, body):
    root = ElementTree.fromstring(body)
    result.worm_id = _find_tag(root, "WormId")
    result.state = _find_tag(root, "State")
    result.retention_period_days = _find_int(root, "RetentionPeriodInDays")
    result.creation_date = _find_tag(root, "CreationDate")

def to_put_extend_bucket_worm(retention_period_days):
    root = ElementTree.Element('ExtendWormConfiguration')
    _add_text_child(root, 'RetentionPeriodInDays', str(retention_period_days))
    return _node_to_string(root)

def to_put_init_bucket_worm(retention_period_days):
    root = ElementTree.Element('InitiateWormConfiguration')
    _add_text_child(root, 'RetentionPeriodInDays', str(retention_period_days))
    return _node_to_string(root)

def to_put_bucket_replication(replication_config):
    root = ElementTree.Element('ReplicationConfiguration')
    rule = ElementTree.SubElement(root, 'Rule')
    if replication_config.rule_id:
        _add_text_child(rule, 'ID', replication_config.rule_id)

    destination = ElementTree.SubElement(rule, 'Destination')
    _add_text_child(destination, 'Bucket', replication_config.target_bucket_name)
    _add_text_child(destination, 'Location', replication_config.target_bucket_location)

    if replication_config.target_transfer_type:
        _add_text_child(destination, 'TransferType', replication_config.target_transfer_type)

    if replication_config.is_enable_historical_object_replication is False:
        _add_text_child(rule, 'HistoricalObjectReplication', 'disabled')
    else:
        _add_text_child(rule, 'HistoricalObjectReplication', 'enabled')

    if replication_config.prefix_list:
        prefix_list_node = ElementTree.SubElement(rule, 'PrefixSet')
        for prefix in replication_config.prefix_list:
            _add_text_child(prefix_list_node, 'Prefix', prefix)

    if replication_config.action_list:
        actions = ''
        for action in replication_config.action_list:
            actions += action
            actions += ','
        actions = actions[:-1]
        _add_text_child(rule, 'Action', actions)

    if replication_config.sync_role_name:
        _add_text_child(rule, 'SyncRole', replication_config.sync_role_name)

    if replication_config.replica_kms_keyid:
        encryption_config = ElementTree.SubElement(rule, 'EncryptionConfiguration')
        _add_text_child(encryption_config, 'ReplicaKmsKeyID', replication_config.replica_kms_keyid)

    if replication_config.sse_kms_encrypted_objects_status in ['Enabled', 'Disabled']:
        criteria = ElementTree.SubElement(rule, 'SourceSelectionCriteria')
        sse_kms_encrypted_objects = ElementTree.SubElement(criteria, 'SseKmsEncryptedObjects')
        _add_text_child(sse_kms_encrypted_objects, 'Status', replication_config.sse_kms_encrypted_objects_status)

    return _node_to_string(root)

def to_delete_bucket_replication(rule_id):
    root = ElementTree.Element('ReplicationRules')
    _add_text_child(root, 'ID', rule_id)

    return _node_to_string(root)

def parse_get_bucket_replication_result(result, body):
    root = ElementTree.fromstring(body)

    for rule_node in root.findall("Rule"):
        rule = ReplicationRule()
        if rule_node.find("ID") is not None:
            rule.rule_id = _find_tag(rule_node, "ID")

        destination_node = rule_node.find("Destination")
        rule.target_bucket_name = _find_tag_with_default(destination_node, "Bucket", None)
        rule.target_bucket_location = _find_tag_with_default(destination_node, "Location", None)
        rule.target_transfer_type = _find_tag_with_default(destination_node, "TransferType", None)
        rule.target_cloud = _find_tag_with_default(destination_node, "Cloud", None)
        rule.target_cloud_location = _find_tag_with_default(destination_node, "CloudLocation", None)

        rule.status = _find_tag(rule_node, "Status")
        rule.sync_role_name = _find_tag_with_default(rule_node, 'SyncRole', None)
        rule.replica_kms_keyid = _find_tag_with_default(rule_node, 'EncryptionConfiguration/ReplicaKmsKeyID', None)
        rule.sse_kms_encrypted_objects_status = _find_tag_with_default(rule_node, 'SourceSelectionCriteria/SseKmsEncryptedObjects/Status', None)

        if _find_tag(rule_node, "HistoricalObjectReplication") == 'enabled':
            rule.is_enable_historical_object_replication = True
        else:
            rule.is_enable_historical_object_replication = False

        prefixes_node = rule_node.find('PrefixSet')
        if prefixes_node is not None:
            rule.prefix_list = _find_all_tags(prefixes_node, 'Prefix')

        actions = _find_tag(rule_node, 'Action')
        rule.action_list = actions.split(',')

        result.rule_list.append(rule)

def parse_get_bucket_replication_location_result(result, body):
    root = ElementTree.fromstring(body)
    result.location_list = _find_all_tags(root, "Location")

    if root.find("LocationTransferTypeConstraint") is not None:
        constraint_node = root.find("LocationTransferTypeConstraint")
        for transfer_type_node in constraint_node.findall("LocationTransferType"):
            location_transfer_type = LocationTransferType()
            location_transfer_type.location = _find_tag_with_default(transfer_type_node, "Location", None)
            location_transfer_type.transfer_type = _find_tag_with_default(transfer_type_node, "TransferTypes/Type", None)
            result.location_transfer_type_list.append(location_transfer_type)

def parse_get_bucket_replication_progress_result(result, body):
    root = ElementTree.fromstring(body)

    rule_node = root.find("Rule")
    progress = BucketReplicationProgress()
    progress.rule_id = _find_tag(rule_node, "ID")

    destination_node = rule_node.find("Destination")
    progress.target_bucket_name = _find_tag(destination_node, "Bucket")
    progress.target_bucket_location = _find_tag(destination_node, "Location")
    progress.target_transfer_type = _find_tag_with_default(destination_node, "TransferType", None)

    progress.status = _find_tag(rule_node, "Status")

    if _find_tag(rule_node, "HistoricalObjectReplication") == 'enabled':
        progress.is_enable_historical_object_replication = True
    else:
        progress.is_enable_historical_object_replication = False

    prefixes_node = rule_node.find('PrefixSet')
    if prefixes_node is not None:
        progress.prefix_list = _find_all_tags(prefixes_node, 'Prefix')

    actions = _find_tag(rule_node, 'Action')
    progress.action_list = actions.split(',')

    historical_object_progress = _find_tag_with_default(rule_node, 'Progress/HistoricalObject', None)
    if historical_object_progress is not None:
        progress.historical_object_progress = float(historical_object_progress)
    progress.new_object_progress = _find_tag_with_default(rule_node, 'Progress/NewObject', None)

    result.progress = progress


def to_put_bucket_transfer_acceleration(enabled):
    root = ElementTree.Element('TransferAccelerationConfiguration')
    _add_text_child(root, 'Enabled', str(enabled))
    return _node_to_string(root)


def parse_get_bucket_transfer_acceleration_result(result, body):
    root = ElementTree.fromstring(body)
    result.enabled = _find_tag(root, "Enabled")


def to_bucket_cname_configuration(domain, cert=None):
    root = ElementTree.Element("BucketCnameConfiguration")
    cname = ElementTree.SubElement(root, 'Cname')
    _add_text_child(cname, 'Domain', domain)
    if cert is not None:
        certificate = ElementTree.SubElement(cname, 'CertificateConfiguration')
        if cert.cert_id is not None:
            _add_text_child(certificate, 'CertId', cert.cert_id)
        if cert.certificate is not None:
            _add_text_child(certificate, 'Certificate', cert.certificate)
        if cert.private_key is not None:
            _add_text_child(certificate, 'PrivateKey',cert.private_key)
        if cert.previous_cert_id is not None:
            _add_text_child(certificate, 'PreviousCertId', cert.previous_cert_id)
        if cert.force is not None:
            _add_text_child(certificate, 'Force', str(cert.force))
        if cert.delete_certificate is not None:
            _add_text_child(certificate, 'DeleteCertificate', str(cert.delete_certificate))
    return _node_to_string(root)


def parse_create_bucket_cname_token(result, body):
    root = ElementTree.fromstring(body)
    result.bucket = _find_tag(root, "Bucket")
    result.cname = _find_tag(root, "Cname")
    result.token = _find_tag(root, "Token")
    result.expire_time = _find_tag(root, "ExpireTime")


def parse_get_bucket_cname_token(result, body):
    root = ElementTree.fromstring(body)
    result.bucket = _find_tag(root, "Bucket")
    result.cname = _find_tag(root, "Cname")
    result.token = _find_tag(root, "Token")
    result.expire_time = _find_tag(root, "ExpireTime")


def parse_list_bucket_cname(result, body):
    root = ElementTree.fromstring(body)
    result.bucket = _find_tag(root, "Bucket")
    result.owner = _find_tag(root, "Owner")
    for cname in root.findall('Cname'):
        tmp = CnameInfo()
        tmp.domain = _find_tag_with_default(cname, 'Domain', None)
        tmp.last_modified = _find_tag_with_default(cname, 'LastModified', None)
        tmp.status = _find_tag_with_default(cname, 'Status', None)
        tmp.is_purge_cdn_cache = _find_tag_with_default(cname, 'IsPurgeCdnCache', None)

        cert = cname.find('Certificate')
        if cert is not None:
            certificate = CertificateInfo()
            certificate.type = _find_tag_with_default(cert, 'Type', None)
            certificate.cert_id = _find_tag_with_default(cert, 'CertId', None)
            certificate.status = _find_tag_with_default(cert, 'Status', None)
            certificate.creation_date = _find_tag_with_default(cert, 'CreationDate', None)
            certificate.fingerprint = _find_tag_with_default(cert, 'Fingerprint', None)
            certificate.valid_start_date = _find_tag_with_default(cert, 'ValidStartDate', None)
            certificate.valid_end_date = _find_tag_with_default(cert, 'ValidEndDate', None)
            tmp.certificate = certificate
        result.cname.append(tmp)



def to_do_bucket_meta_query_request(meta_query):
    root = ElementTree.Element("MetaQuery")
    if meta_query.next_token is not None:
        _add_text_child(root, "NextToken", meta_query.next_token)
    _add_text_child(root, "MaxResults", meta_query.max_results)
    _add_text_child(root, "Query", meta_query.query)
    if meta_query.sort is not None:
        _add_text_child(root, "Sort", meta_query.sort)
    if meta_query.order is not None:
        _add_text_child(root, "Order", meta_query.order)
    if meta_query.aggregations:
        aggregations_node = ElementTree.SubElement(root, "Aggregations")
        for aggregation in meta_query.aggregations:
            aggregation_node = ElementTree.SubElement(aggregations_node, 'Aggregation')
            if aggregation.field is not None:
                _add_text_child(aggregation_node, 'Field', aggregation.field)
            if aggregation.operation is not None:
                _add_text_child(aggregation_node, 'Operation', aggregation.operation)

    return _node_to_string(root)


def parse_get_bucket_meta_query_result(result, body):
    root = ElementTree.fromstring(body)
    result.state = _find_tag(root, "State")
    result.phase = _find_tag(root, "Phase")
    result.create_time = _find_tag(root, "CreateTime")
    result.update_time = _find_tag(root, "UpdateTime")


def parse_do_bucket_meta_query_result(result, body):
    root = ElementTree.fromstring(body)
    result.next_token = _find_tag(root, "NextToken")

    for file in root.findall('Files/File'):
        tmp = MetaQueryFile()
        tmp.file_name = _find_tag(file, 'Filename')
        tmp.size = int(_find_tag_with_default(file, 'Size', 0))
        tmp.file_modified_time = _find_tag_with_default(file, 'FileModifiedTime', None)
        tmp.file_create_time = _find_tag_with_default(file, 'FileCreateTime', None)
        tmp.file_access_time = _find_tag_with_default(file, 'FileAccessTime', None)
        tmp.oss_object_type = _find_tag_with_default(file, 'OSSObjectType', None)
        tmp.oss_storage_class = _find_tag_with_default(file, 'OSSStorageClass', None)
        tmp.object_acl = _find_tag_with_default(file, 'ObjectACL', None)
        tmp.etag = _find_tag_with_default(file, 'ETag', None)
        tmp.oss_crc64 = _find_tag_with_default(file, 'OSSCRC64', None)
        tmp.oss_tagging_count = int(_find_tag_with_default(file, 'OSSTaggingCount', 0))
        if file.find('OSSTagging') is not None:
            for tagging in file.find('OSSTagging').findall('Tagging'):
                tmp_tagging = OSSTaggingInfo(_find_tag(tagging, 'Key'), _find_tag(tagging, 'Value'))
                tmp.oss_tagging.append(tmp_tagging)
        if file.find('OSSUserMeta') is not None:
            for meta in file.find('OSSUserMeta').findall('UserMeta'):
                tmp_meta = OSSUserMetaInfo(_find_tag(meta, 'Key'), _find_tag(meta, 'Value'))
                tmp.oss_user_meta.append(tmp_meta)
        result.files.append(tmp)

    for aggregation in root.findall('Aggregations/Aggregation'):
        tmp = AggregationsInfo()
        tmp.field = _find_tag(aggregation, 'Field')
        tmp.operation = _find_tag(aggregation, 'Operation')
        tmp.value = float(_find_tag_with_default(aggregation, 'Value', 0))

        for group in aggregation.findall('Groups/Group'):
            tmp_groups = AggregationGroupInfo(_find_tag(group, 'Value'), int(_find_tag_with_default(group, 'Count', 0)))
            tmp.groups.append(tmp_groups)
        result.aggregations.append(tmp)

def parse_dummy_result(result, body):
    return result

def to_put_bucket_access_monitor(status):
    root = ElementTree.Element('AccessMonitorConfiguration')
    _add_text_child(root, 'Status', status)
    return _node_to_string(root)

def parse_get_bucket_access_monitor_result(result, body):
    root = ElementTree.fromstring(body)

    access_monitor = AccessMonitorInfo(_find_tag(root, "Status"))
    result.access_monitor = access_monitor

def parse_lifecycle_filter(filter_node):
    lifecycle_filter = LifecycleFilter()
    if filter_node is not None:
        if filter_node.find('ObjectSizeGreaterThan') is not None:
            lifecycle_filter.object_size_greater_than = int(_find_tag_with_default(filter_node, 'ObjectSizeGreaterThan', 0))
        if filter_node.find('ObjectSizeLessThan') is not None:
            lifecycle_filter.object_size_less_than = int(_find_tag_with_default(filter_node, 'ObjectSizeLessThan', 0))
        not_nodes = filter_node.findall('Not')
        if not_nodes is not None:
            for not_node in not_nodes:
                prefix = _find_tag_with_default(not_node, 'Prefix', None)
                key = _find_tag_with_default(not_node, 'Tag/Key', None)
                value = _find_tag_with_default(not_node, 'Tag/Value', None)
                tag = FilterNotTag(key, value)
                filter_not = FilterNot(prefix, tag)
                lifecycle_filter.filter_not.append(filter_not)

    return lifecycle_filter

def to_put_bucket_resource_group(resourceGroupId):
    root = ElementTree.Element('BucketResourceGroupConfiguration')
    _add_text_child(root, 'ResourceGroupId', resourceGroupId)
    return _node_to_string(root)

def parse_get_bucket_resource_group_result(result, body):
    root = ElementTree.fromstring(body)
    result.resource_group_id = _find_tag(root, "ResourceGroupId")

def to_put_bucket_style(content):
    root = ElementTree.Element('Style')
    _add_text_child(root, 'Content', content)
    return _node_to_string(root)

def parse_get_bucket_style_result(result, body):
    root = ElementTree.fromstring(body)
    result.name = _find_tag(root, "Name")
    result.content = _find_tag(root, "Content")
    result.create_time = _find_tag(root, "CreateTime")
    result.last_modify_time = _find_tag(root, "LastModifyTime")

def parse_list_bucket_style(result, body):
    root = ElementTree.fromstring(body)
    for style in root.findall('Style'):
        tmp = BucketStyleInfo()
        tmp.name = _find_tag_with_default(style, 'Name', None)
        tmp.content = _find_tag_with_default(style, 'Content', None)
        tmp.create_time = _find_tag_with_default(style, 'CreateTime', None)
        tmp.last_modify_time = _find_tag_with_default(style, 'LastModifyTime', None)

        result.styles.append(tmp)


def parse_describe_regions(result, body):
    root = ElementTree.fromstring(body)
    for region in root.findall('RegionInfo'):
        tmp = RegionInfo()
        tmp.region = _find_tag_with_default(region, 'Region', None)
        tmp.internet_endpoint = _find_tag_with_default(region, 'InternetEndpoint', None)
        tmp.internal_endpoint = _find_tag_with_default(region, 'InternalEndpoint', None)
        tmp.accelerate_endpoint = _find_tag_with_default(region, 'AccelerateEndpoint', None)

        result.regions.append(tmp)

def parse_async_process_object(result, body):
    if body:
        body_dict = json.loads(to_unicode(body))
        result.event_id = body_dict['EventId']
        result.async_request_id = body_dict['RequestId']
        result.task_id = body_dict['TaskId']
    return result

def to_do_bucket_callback_policy_request(callback_policy):
    root = ElementTree.Element("BucketCallbackPolicy")
    if callback_policy:
        for policy in callback_policy:
            if policy:
                policy_node = ElementTree.SubElement(root, 'PolicyItem')
                if policy.policy_name is not None:
                    _add_text_child(policy_node, 'PolicyName', policy.policy_name)
                if policy.callback is not None:
                    _add_text_child(policy_node, 'Callback', policy.callback)
                if policy.callback_var is None:
                    _add_text_child(policy_node, 'CallbackVar', '')
                else:
                    _add_text_child(policy_node, 'CallbackVar', policy.callback_var)

    return _node_to_string(root)

def parse_callback_policy_result(result, body):
    root = ElementTree.fromstring(body)
    for policy in root.findall('PolicyItem'):
        tmp = CallbackPolicyInfo()
        tmp.policy_name = _find_tag_with_default(policy, 'PolicyName', None)
        tmp.callback = _find_tag_with_default(policy, 'Callback', None)
        tmp.callback_var = _find_tag_with_default(policy, 'CallbackVar', None)

        result.callback_policies.append(tmp)

def to_put_bucket_archive_direct_read(enabled):
    root = ElementTree.Element('ArchiveDirectReadConfiguration')
    _add_text_child(root, 'Enabled', str(enabled).lower())
    return _node_to_string(root)

def parse_get_bucket_archive_direct_read(result, body):
    root = ElementTree.fromstring(body)
    if root.find("Enabled") is not None:
        result.enabled = _find_bool(root, "Enabled")
        
def to_do_bucket_https_config_request(https_config):
    root = ElementTree.Element('HttpsConfiguration')

    list_node = ElementTree.SubElement(root, 'TLS')

    _add_text_child(list_node, 'Enable', str(https_config.tls_enabled).lower())
    if https_config.tls_version:
        for r in https_config.tls_version:
            _add_text_child(list_node, 'TLSVersion', r)

    return _node_to_string(root)


def parse_get_bucket_https_config(result, body):
    root = ElementTree.fromstring(body)

    result.tls_enabled = _find_bool(root, 'TLS/Enable')
    if root.find("TLS/TLSVersion") is not None:
        result.tls_version = _find_all_tags(root, 'TLS/TLSVersion')

    return result


def parse_create_data_redundancy_transition_result(result, body):
    root = ElementTree.fromstring(body)
    result.task_id = _find_tag_with_default(root, "TaskId", None)

def parse_data_redundancy_transition(transition_node):
    if transition_node is None:
        return None

    transition = DataRedundancyTransitionInfo()

    if transition_node.find('Bucket') is not None:
        transition.bucket = _find_tag_with_default(transition_node, 'Bucket', None)
    if transition_node.find('TaskId') is not None:
        transition.task_id = _find_tag_with_default(transition_node, 'TaskId', None)
    if transition_node.find('Status') is not None:
        transition.transition_status = _find_tag_with_default(transition_node, 'Status', None)
    if transition_node.find('CreateTime') is not None:
        transition.create_time = _find_tag_with_default(transition_node, 'CreateTime', None)
    if transition_node.find('StartTime') is not None:
        transition.start_time = _find_tag_with_default(transition_node, 'StartTime', None)
    if transition_node.find('EndTime') is not None:
        transition.end_time = _find_tag_with_default(transition_node, 'EndTime', None)
    if transition_node.find('EstimatedRemainingTime') is not None:
        transition.estimated_remaining_time = _find_int(transition_node, 'EstimatedRemainingTime')
    if transition_node.find('ProcessPercentage') is not None:
        transition.process_percentage = _find_int(transition_node, 'ProcessPercentage')

    return transition


def parse_get_bucket_data_redundancy_transition(result, body):
    root = ElementTree.fromstring(body)
    if root.find('Bucket') is not None:
        result.bucket = _find_tag_with_default(root, 'Bucket', None)
    if root.find('TaskId') is not None:
        result.task_id = _find_tag_with_default(root, 'TaskId', None)
    if root.find('Status') is not None:
        result.transition_status = _find_tag_with_default(root, 'Status', None)
    if root.find('CreateTime') is not None:
        result.create_time = _find_tag_with_default(root, 'CreateTime', None)
    if root.find('StartTime') is not None:
        result.start_time = _find_tag_with_default(root, 'StartTime', None)
    if root.find('EndTime') is not None:
        result.end_time = _find_tag_with_default(root, 'EndTime', None)
    if root.find('EstimatedRemainingTime') is not None:
        result.estimated_remaining_time = _find_int(root, 'EstimatedRemainingTime')
    if root.find('ProcessPercentage') is not None:
        result.process_percentage = _find_int(root, 'ProcessPercentage')


def parse_list_bucket_data_redundancy_transition(result, body):
    root = ElementTree.fromstring(body)
    for transition in root.findall('BucketDataRedundancyTransition'):
        tmp = parse_data_redundancy_transition(transition)
        result.data_redundancy_transitions.append(tmp)

def parse_list_user_data_redundancy_transition(result, body):
    root = ElementTree.fromstring(body)
    if root.find("IsTruncated") is not None:
        result.is_truncated = _find_bool(root, "IsTruncated")

    if root.find("NextContinuationToken") is not None:
        result.next_continuation_token = _find_tag(root, "NextContinuationToken")

    for transition in root.findall('BucketDataRedundancyTransition'):
        tmp = parse_data_redundancy_transition(transition)
        result.data_redundancy_transitions.append(tmp)

def to_do_create_access_point_request(accessPoint):
    root = ElementTree.Element("CreateAccessPointConfiguration")
    if accessPoint.access_point_name is not None:
        _add_text_child(root, "AccessPointName", accessPoint.access_point_name)
    if accessPoint.network_origin is not None:
        _add_text_child(root, "NetworkOrigin", accessPoint.network_origin)
    if accessPoint.vpc is not None:
        vpc_node = ElementTree.SubElement(root, "VpcConfiguration")
        _add_text_child(vpc_node, "VpcId", accessPoint.vpc.vpc_id)

    return _node_to_string(root)


def parse_create_access_point_result(result, body):
    root = ElementTree.fromstring(body)
    result.access_point_arn = _find_tag_with_default(root, "AccessPointArn", None)
    result.alias = _find_tag_with_default(root, "Alias", None)


def parse_get_access_point_result(result, body):
    root = ElementTree.fromstring(body)
    result.access_point_name = _find_tag_with_default(root, "AccessPointName", None)
    result.bucket = _find_tag_with_default(root, "Bucket", None)
    result.account_id = _find_tag_with_default(root, "AccountId", None)
    result.network_origin = _find_tag_with_default(root, "NetworkOrigin", None)
    vpc_node = root.find('VpcConfiguration')
    if vpc_node is not None:
        vpc = AccessPointVpcConfiguration()
        vpc.vpc_id = _find_tag_with_default(vpc_node, "VpcId", None)
        result.vpc = vpc

    result.access_point_arn = _find_tag_with_default(root, "AccessPointArn", None)
    result.creation_date = _find_tag_with_default(root, "CreationDate", None)
    result.alias = _find_tag_with_default(root, "Alias", None)
    result.access_point_status = _find_tag_with_default(root, "Status", None)
    endpoint_node = root.find('Endpoints')
    if endpoint_node is not None:
        endpoint = AccessPointEndpoints()
        endpoint.public_endpoint = _find_tag_with_default(endpoint_node, "PublicEndpoint", None)
        endpoint.internal_endpoint = _find_tag_with_default(endpoint_node, "InternalEndpoint", None)
        result.endpoints = endpoint

    block_node = root.find('PublicAccessBlockConfiguration')
    if block_node is not None:
        block = PublicAccessBlockConfiguration()
        if block_node.find('BlockPublicAccess') is not None:
            block.block_public_access = _find_bool(block_node, "BlockPublicAccess")
            result.public_access_block_configuration = block


def parse_list_access_point_result(result, body):
    root = ElementTree.fromstring(body)

    result.account_id = _find_tag_with_default(root, 'AccountId', None)
    result.marker = _find_tag_with_default(root, 'Marker', None)
    result.max_keys = _find_int(root, 'MaxKeys')
    result.is_truncated = _find_bool(root, 'IsTruncated')

    if result.is_truncated is not None:
        result.next_continuation_token = _find_tag_with_default(root, 'NextContinuationToken', None)

    access_points = root.findall('AccessPoints/AccessPoint')
    for access_point in access_points:
        tmp = AccessPointInfo()
        tmp.bucket = _find_tag(access_point, 'Bucket')
        tmp.access_point_name = _find_tag(access_point, 'AccessPointName')
        tmp.alias = _find_tag(access_point, 'Alias')
        tmp.status = _find_tag(access_point, 'Status')
        tmp.network_origin = _find_tag(access_point, 'NetworkOrigin')
        vpc_node = access_point.find('VpcConfiguration')
        if vpc_node is not None:
            vpc = AccessPointVpcConfiguration()
            vpc.vpc_id = _find_tag_with_default(vpc_node, "VpcId", None)
            tmp.vpc = vpc
        result.access_points.append(tmp)
    return result


def to_put_public_access_block_request(block_public_access):
    root = ElementTree.Element('PublicAccessBlockConfiguration')
    _add_text_child(root, 'BlockPublicAccess', str(block_public_access))
    return _node_to_string(root)


def parse_get_public_access_block_result(result, body):
    root = ElementTree.fromstring(body)
    if root.find("BlockPublicAccess") is not None:
        result.block_public_access = _find_bool(root, 'BlockPublicAccess')


def parse_qos_configuration(qos_info_node):
    if qos_info_node is None:
        return None

    qos = QoSConfiguration()
    if qos_info_node.find('TotalUploadBandwidth') is not None:
        qos.total_upload_bw = _find_int(qos_info_node, 'TotalUploadBandwidth')
    if qos_info_node.find('IntranetUploadBandwidth') is not None:
        qos.intranet_upload_bw = _find_int(qos_info_node, 'IntranetUploadBandwidth')
    if qos_info_node.find('ExtranetUploadBandwidth') is not None:
        qos.extranet_upload_bw = _find_int(qos_info_node, 'ExtranetUploadBandwidth')
    if qos_info_node.find('TotalDownloadBandwidth') is not None:
        qos.total_download_bw = _find_int(qos_info_node, 'TotalDownloadBandwidth')
    if qos_info_node.find('IntranetDownloadBandwidth') is not None:
        qos.intranet_download_bw = _find_int(qos_info_node, 'IntranetDownloadBandwidth')
    if qos_info_node.find('ExtranetDownloadBandwidth') is not None:
        qos.extranet_download_bw = _find_int(qos_info_node, 'ExtranetDownloadBandwidth')
    if qos_info_node.find('TotalQps') is not None:
        qos.total_qps = _find_int(qos_info_node, 'TotalQps')
    if qos_info_node.find('IntranetQps') is not None:
        qos.intranet_qps = _find_int(qos_info_node, 'IntranetQps')
    if qos_info_node.find('ExtranetQps') is not None:
        qos.extranet_qps = _find_int(qos_info_node, 'ExtranetQps')

    return qos


def to_put_bucket_requester_qos_info(enabled):
    root = ElementTree.Element('QoSConfiguration')
    _add_text_child(root, 'Enabled', str(enabled).lower())
    return _node_to_string(root)


def parse_get_requester_qos_info(result, body):
    root = ElementTree.fromstring(body)

    result.requester = _find_tag_with_default(root, 'Requester', None)
    result.qos_configuration = parse_qos_configuration(root.find('QoSConfiguration'))

    return result

def parse_get_resource_pool_info(result, body):
    root = ElementTree.fromstring(body)

    result.region = _find_tag_with_default(root, 'Region', None)
    result.name = _find_tag_with_default(root, 'Name', None)
    result.owner = _find_tag_with_default(root, 'Owner', None)
    result.create_time = _find_tag_with_default(root, 'CreateTime', None)
    result.qos_configuration = parse_qos_configuration(root.find('QoSConfiguration'))

    return result


def parse_list_resource_pools(result, body):
    root = ElementTree.fromstring(body)

    result.region = _find_tag_with_default(root, 'Region', None)
    result.owner = _find_tag_with_default(root, 'Owner', None)
    result.continuation_token = _find_tag_with_default(root, 'ContinuationToken', None)
    result.next_continuation_token = _find_tag_with_default(root, 'NextContinuationToken', None)
    result.is_truncated = _find_bool(root, 'IsTruncated')

    for resource in root.findall('ResourcePool'):
        tmp = ResourcePoolInfo()
        tmp.name = _find_tag_with_default(resource, 'Name', None)
        tmp.create_time = _find_tag_with_default(resource, 'CreateTime', None)
        result.resource_pool.append(tmp)

    return result


def parse_list_resource_pool_buckets(result, body):
    root = ElementTree.fromstring(body)

    result.resource_pool = _find_tag_with_default(root, 'ResourcePool', None)
    result.continuation_token = _find_tag_with_default(root, 'ContinuationToken', None)
    result.next_continuation_token = _find_tag_with_default(root, 'NextContinuationToken', None)
    result.is_truncated = _find_bool(root, 'IsTruncated')

    for bucket in root.findall('ResourcePoolBucket'):
        tmp = ResourcePoolBucketInfo()
        tmp.name = _find_tag_with_default(bucket, 'Name', None)
        tmp.join_time = _find_tag_with_default(bucket, 'JoinTime', None)

        result.resource_pool_buckets.append(tmp)

    return result


def parse_list_resource_pool_requester_qos_infos(result, body):
    root = ElementTree.fromstring(body)

    result.resource_pool = _find_tag_with_default(root, 'ResourcePool', None)
    result.continuation_token = _find_tag_with_default(root, 'ContinuationToken', None)
    result.next_continuation_token = _find_tag_with_default(root, 'NextContinuationToken', None)
    result.is_truncated = _find_bool(root, 'IsTruncated')

    for qos_info in root.findall('RequesterQoSInfo'):
        tmp = RequesterQoSInfo()
        tmp.requester = _find_tag_with_default(qos_info, 'Requester', None)
        tmp.qos_configuration = parse_qos_configuration(qos_info.find('QoSConfiguration'))

        result.requester_qos_info.append(tmp)

    return result


def parse_list_bucket_requester_qos_infos(result, body):
    root = ElementTree.fromstring(body)

    result.bucket = _find_tag_with_default(root, 'Bucket', None)
    result.continuation_token = _find_tag_with_default(root, 'ContinuationToken', None)
    result.next_continuation_token = _find_tag_with_default(root, 'NextContinuationToken', None)
    result.is_truncated = _find_bool(root, 'IsTruncated')

    for qos_info in root.findall('RequesterQoSInfo'):
        tmp = RequesterQoSInfo()
        tmp.requester = _find_tag_with_default(qos_info, 'Requester', None)
        tmp.qos_configuration = parse_qos_configuration(qos_info.find('QoSConfiguration'))

        result.requester_qos_info.append(tmp)
    return result
