"""EDOS PDS construction record generation for SUOMI NPP"""

from datetime import datetime
import itertools
import os
from edosl0util.headers import DaySegmentedTimecode
from edosl0util.stream import jpss_packet_stream

def test_build_apid_info():
    # FIXME: build CR comparison into the CLI
    calculated = build_cr('P1571289CRISSCIENCEAAT15320210920101.PDS')
    ingested = read_pds_cr('P1571289CRISSCIENCEAAT15320210920100.PDS')
    insert_fake_cr_info(ingested)
    del calculated['completion_time']  # it seems CR completion time does not match PDS
    del ingested['completion_time']    # creation time from the file name
    assert calculated == ingested

def build_cr(pds_file, prev_pds_file=None):
    """Best-effort CR generation by scanning a PDS data file

    Previous PDS data file may also be given to make gap detection more complete.
    """
    def main():
        scan = scan_packets(pds_file, prev_pds_file)
        rv = {}
        rv['pds_id'] = pds_id_from_path(pds_file)
        rv['completion_time'] = get_pds_creation_time(pds_file)
        rv['first_packet_time'] = scan['first_packet_time']
        rv['last_packet_time'] = scan['last_packet_time']
        rv['apid_info'] = build_apid_info(scan['apid_info'])
        rv['apid_count'] = len(rv['apid_info'])
        rv['file_info'] = file_info(rv['pds_id'], rv['apid_info'])
        rv['file_count'] = len(rv['file_info'])
        rv.update(aggregated_values(rv['apid_info']))
        insert_fake_cr_info(rv)
        return rv
    def file_info(pds_id, apid_info):
        cr_entry = {'file_name': pds_id + '.PDS', 'apid_count': 0, 'apid_info' : []}
        data_entry = {'file_name': pds_file,
                      'apid_count': len(apid_info),
                      'apid_info': [{'scid': npp_scid, 'apid': entry['apid'],
                                     'first_packet_time': entry['first_packet_time'],
                                     'last_packet_time': entry['last_packet_time']}
                                        for entry in apid_info]}
        return [cr_entry, data_entry]
    def aggregated_values(apid_info):
        keys = ['total_packets', 'total_bytes', 'gap_count', 'fill_bytes',
                'mismatched_length_packets', 'rs_corrected_packets']
        return {key: sum(entry[key] for entry in apid_info) for key in keys}
    return main()

def insert_fake_cr_info(cr):
    """Populate a CR with phony values for fields that can't be discovered via a packet scan"""
    cr.update({'edos_sw_ver_major': 0, 'edos_sw_ver_minor': 0, 'cr_type': 1, 'test_flag': 0,
               'scs_count': 1, 'scs_info': {'start': missing_time_value,
                                             'stop': missing_time_value},
               'first_packet_esh_time': missing_time_value,
               'last_packet_esh_time': missing_time_value,
               'fill_bytes': 0, 'mismatched_length_packets': 0, 'rs_corrected_packets': 0})
    insert_fake_apid_info(cr['apid_info'])

def pds_id_from_path(pds_file):
    """Pull 36-char PDS ID from a file name; that's the CR file name minus the .PDS"""
    file_name = os.path.basename(pds_file)
    pds_file_name_length = 40
    if len(file_name) != pds_file_name_length:
        raise ValueError('PDS file name {} not of expected length {}'.format(
                             file_name, pds_file_name_length))
    return file_name[:34] + '00'

def get_pds_creation_time(pds_file_or_id):
    """Parse 11-char creation time out of a PDS ID or file name; return a DaySegmentedTimecode"""
    return datetime_to_ccsds(datetime.strptime(pds_file_or_id[22:33], '%y%j%H%M%S'))

def build_apid_info(scan_apid_info):
    """Build up apid_info resulting from scan_packets into a full apid_info for a CR"""
    for entry in scan_apid_info:
        entry['scid'] = npp_scid
        entry['vcid_count'] = 1
        entry['vcid_info'] = [{'scid': npp_scid, 'vcid': npp_apid_to_vcid_map[entry['apid']]}]
        entry['gap_count'] = len(entry['gap_info'])
        del entry['last_packet_ssc']  # field was needed for bookkeeping but isn't in the CR
    insert_fake_apid_info(apid_info)
    return apid_info

def insert_fake_apid_info(apid_info):
    """Fill CR apid_info with phony values for fields that can't be found via a packet scan"""
    for entry in apid_info:
        entry.update({
            'fill_packets': 0, 'fill_packet_info': [], 'fill_bytes': 0,
            'mismatched_length_packets': 0, 'mismatched_length_packet_ssc_list': [],
            'first_packet_esh_time': missing_time_value,
            'last_packet_esh_time': missing_time_value,
            'rs_corrected_packets': 0})
        for gap in entry['gap_info']:
            gap.update({'pre_gap_packet_esh_time': missing_time_value,
                        'post_gap_packet_esh_time': missing_time_value})

missing_time_value = DaySegmentedTimecode(0, 0, 0)

npp_scid = 157  # FIXME: i guess this should come from the file name

# ripped from MDFCB, 2014-06-05 revision
# modified to place CrIS FOV 6 in VCID 7 after testing against some real data
npp_vcid_to_apids_map = {
    0: list(range(0, 15)) + list(range(16, 50)) + [65, 70, 100, 146, 155, 512, 513, 518,
       543, 544, 545, 550, 768, 769, 773] + list(range(1280, 1289)),
    1: [50, 101, 515, 528, 530, 531],
    2: [102],
    3: [103, 514, 536],
    4: [104],
    5: [105],
    6: [106, 1289, 1290] + list(set(range(1315, 1396)) - set(range(1318, 1391, 9))
                                    - set(range(1320, 1393, 9))),
    7: [107] + list(range(1318, 1391, 9)) + list(range(1320, 1393, 9)),
    8: [108, 1294, 1295, 1296, 1398],
    9: [109],
    10: [110],
    11: [111, 560, 561, 564, 565, 576],
    12: [112, 562, 563, 566],
    13: [113, 546, 577, 578, 579, 580, 581, 582],
    14: [114],
    15: [115],
    16: [116] + list(range(800, 806)) + list(range(807, 824)) + [825, 826],
    17: [117, 806],
    18: [118, 770] + list(range(830, 854)) + [855, 856],
    19: [119],
    20: [120],
    21: [121, 517, 524, 549, 556, 780, 1291, 1292, 1293, 1397],
    22: [122],
    24: [147, 148, 149, 150]}
npp_apid_to_vcid_map = {apid: vcid for vcid, apids in npp_vcid_to_apids_map.items()
                                   for apid in apids}

def scan_packets(pds_file, prev_pds_file=None):
    """Scan a PDS data file for information needed to produce a construction record"""
    def main():
        prev_apid_map = build_prev_apid_map(prev_pds_file)
        apid_map = {}
        stream = jpss_packet_stream(open(pds_file, 'rb'))
        first_pkt = stream.next()
        for pkt in itertools.chain([first_pkt], stream):
            entry = apid_map.get(pkt.apid)
            if not entry:
                entry_from_prev_pds = prev_apid_map.get(pkt.apid)
                apid_map[pkt.apid] = init_entry(pkt, entry_from_prev_pds)
            else:
                update_entry(entry, pkt)
        last_pkt = pkt
        return {'first_packet_time': datetime_to_ccsds(first_pkt.stamp),
                'last_packet_time': datetime_to_ccsds(last_pkt.stamp),
                'apid_info': [apid_map[k] for k in sorted(apid_map)]}
    def build_prev_apid_map(prev_pds_file):
        if prev_pds_file:
            return {entry['apid']: entry for entry in scan_packets(prev_pds_file)['apid_info']}
        else:
            return {}
    def init_entry(pkt, entry_from_prev_pds):
        rv = {'apid': pkt.apid,
              'first_packet_time': datetime_to_ccsds(pkt.stamp), 'first_packet_offset': pkt.offset,
              'last_packet_time': datetime_to_ccsds(pkt.stamp), 'last_packet_ssc': pkt.seqid,
              'total_packets': 1, 'total_bytes': pkt.size, 'gap_info': []}
        if entry_from_prev_pds:
            update_gap_info(rv['gap_info'], entry_from_prev_pds['last_packet_ssc'],
                            entry_from_prev_pds['last_packet_time'], pkt)
        return rv
    def update_entry(entry, new_pkt):
        prev_last_ssc = entry['last_packet_ssc']
        prev_last_time = entry['last_packet_time']
        entry['last_packet_time'] = datetime_to_ccsds(new_pkt.stamp)
        entry['last_packet_ssc'] = new_pkt.seqid
        entry['total_packets'] += 1
        entry['total_bytes'] += new_pkt.size
        update_gap_info(entry['gap_info'], prev_last_ssc, prev_last_time, new_pkt)
    def update_gap_info(gap_info, last_ssc, last_pkt_time, new_pkt):
        ssc_limit = 16384  # one more than highest SSC
        expected_new_ssc = (last_ssc + 1) % ssc_limit
        if new_pkt.seqid != expected_new_ssc:
            gap_entry = {'first_missing_ssc': expected_new_ssc,
                         'missing_packet_count': (new_pkt.seqid - expected_new_ssc) % ssc_limit,
                         'pre_gap_packet_time': last_pkt_time,
                         'post_gap_packet_time': datetime_to_ccsds(new_pkt.stamp),
                         'post_gap_packet_offset': new_pkt.offset}
            gap_info.append(gap_entry)
    return main()

def datetime_to_ccsds(dt):
    """Convert a packet stamp to DaySegmentedTimecode

    Handles input of None by returning epoch value of 1958-01-01.
    """
    if dt is not None:
        epoch = datetime(1958, 1, 1)
        days = (dt - epoch).days
        micros = int((dt - datetime(dt.year, dt.month, dt.day)).total_seconds() * 1e6)
        return DaySegmentedTimecode(days, micros // 1000, micros % 1000)
    else:
        return DaySegmentedTimecode()