import logging
import json
import time

from datetime import timedelta, datetime
from gzip import GzipFile
from tempfile import SpooledTemporaryFile

import features
from app import app, storage
from data.database import UseThenDisconnect
from data.model.log import (get_stale_logs, get_stale_logs_start_id,
                            get_stale_logs_cutoff_id, delete_stale_logs)
from data.userfiles import DelegateUserfiles
from util.locking import GlobalLock
from util.streamingjsonencoder import StreamingJSONEncoder
from workers.worker import Worker

logger = logging.getLogger(__name__)

JSON_MIMETYPE = 'application/json'
STALE_AFTER = timedelta(days=30)
MIN_LOGS_PER_ROTATION = 10000
MEMORY_TEMPFILE_SIZE = 12 * 1024 * 1024

WORKER_FREQUENCY = app.config.get('ACTION_LOG_ROTATION_FREQUENCY', 60 * 60 * 24)
SAVE_PATH = app.config.get('ACTION_LOG_ARCHIVE_PATH')
SAVE_LOCATION = app.config.get('ACTION_LOG_ARCHIVE_LOCATION')

class LogRotateWorker(Worker):
  """ Worker used to rotate old logs out the database and into storage. """
  def __init__(self):
    super(LogRotateWorker, self).__init__()
    self.add_operation(self._archive_logs, WORKER_FREQUENCY)

  def _archive_logs(self):
    logger.debug('Attempting to rotate log entries')

    log_archive = DelegateUserfiles(app, storage, SAVE_LOCATION, SAVE_PATH)

    with UseThenDisconnect(app.config):
      cutoff_date = datetime.now() - STALE_AFTER
      cutoff_id = get_stale_logs_cutoff_id(cutoff_date)
      if cutoff_id is None:
        logger.warning('Failed to find cutoff id')
        return

    while True:
      with GlobalLock('ACTION_LOG_ROTATION') as gl:
        if not gl:
          logger.debug('Could not acquire global lock; sleeping')
          return

        with UseThenDisconnect(app.config):
          start_id = get_stale_logs_start_id()

          if start_id is None:
            logger.warning('Failed to find start id')
            return

          logger.debug('Found starting ID %s and cutoff ID %s', start_id, cutoff_id)

          approx_count = cutoff_id - start_id
          if approx_count < MIN_LOGS_PER_ROTATION:
            logger.debug('Not enough stale logs to warrant rotation (approx %d)', approx_count)
            return

          end_id = start_id + MIN_LOGS_PER_ROTATION
          logs = [log_dict(log) for log in get_stale_logs(start_id, end_id)]

        logger.debug('Archiving logs from IDs %s to %s', start_id, end_id)
        with SpooledTemporaryFile(MEMORY_TEMPFILE_SIZE) as tempfile:
          with GzipFile('temp_action_log_rotate', fileobj=tempfile, compresslevel=1) as zipstream:
            for chunk in StreamingJSONEncoder().iterencode(logs):
              zipstream.write(chunk)

          tempfile.seek(0)
          filename = '%d-%d.txt.gz' % (start_id, end_id)
          log_archive.store_file(tempfile, JSON_MIMETYPE, content_encoding='gzip',
                                 file_id=filename)
          logger.debug('Finished archiving logs from IDs %s to %s', start_id, end_id)

        with UseThenDisconnect(app.config):
          logger.debug('Deleting logs from IDs %s to %s', start_id, end_id)
          delete_stale_logs(start_id, end_id)


def log_dict(log):
  """ Pretty prints a LogEntry in JSON. """
  return {'kind_id': log.kind_id,
          'account_id': log.account_id,
          'performer_id': log.performer_id,
          'repository_id': log.repository_id,
          'datetime': str(log.datetime),
          'ip': str(log.ip),
          'metadata_json': json.loads(str(log.metadata_json))}


def main():
  logging.config.fileConfig('conf/logging_debug.conf', disable_existing_loggers=False)

  if not features.ACTION_LOG_ROTATION or None in [SAVE_PATH, SAVE_LOCATION]:
    logger.debug('Action log rotation worker not enabled; skipping')
    while True:
      time.sleep(100000)

  worker = LogRotateWorker()
  worker.start()


if __name__ == "__main__":
  main()