summaryrefslogtreecommitdiffstats
path: root/include/framework/scheduling/statistics.h
blob: 6c479cd921ebe544c2d1d8bf2f08321181760a6e (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
/*
 * include/framework/scheduling/statistics.h
 *
 * Copyright (C) 2023 Douglas B. Rumbaugh <drumbaugh@psu.edu> 
 *
 * Distributed under the Modified BSD License.
 *
 * This is a stub for a statistics tracker to be used in scheduling. It
 * currently only tracks simple aggregated statistics, but should be
 * updated in the future for more fine-grained statistics. These will be
 * used for making scheduling decisions and predicting the runtime of a
 * given job.
 */
#pragma once

#include <cstdlib>
#include <cassert>
#include <unordered_map>
#include <vector>
#include <mutex>
#include <chrono>
#include <atomic>

namespace de {

class SchedulerStatistics {
private:
    enum class EventType {
        QUEUED,
        SCHEDULED,
        STARTED, 
        FINISHED
    };

    struct Event {
        size_t id;
        EventType type;
    };

    struct JobInfo {
        size_t id;
        size_t size;
        size_t type;
    };


public:
    SchedulerStatistics() = default;
    ~SchedulerStatistics() = default;

    void job_queued(size_t id, size_t type, size_t size) {
        auto time = std::chrono::high_resolution_clock::now();
    }

    void job_scheduled(size_t id) {
        std::unique_lock<std::mutex> lk(m_mutex);

    }

    void job_begin(size_t id) {

    }

    void job_complete(size_t id) {

    }

    /* FIXME: This is just a temporary approach */
    void log_time_data(size_t length, size_t type) {
        assert(type == 1 || type == 2);

        if (type == 1) {
            m_type_1_cnt.fetch_add(1);
            m_type_1_total_time.fetch_add(length);
            
            if (length > m_type_1_largest_time) {
                m_type_1_largest_time.store(length);
            }
        } else {
            m_type_2_cnt.fetch_add(1);
            m_type_2_total_time.fetch_add(length);

            if (length > m_type_2_largest_time) {
                m_type_2_largest_time.store(length);
            }
        }
    }

    void print_statistics() {
        if (m_type_1_cnt > 0) {
            fprintf(stdout, "Query Count: %ld\tQuery Avg. Latency: %ld\tMax Query Latency: %ld\n", 
                    m_type_1_cnt.load(), 
                    m_type_1_total_time.load() / m_type_1_cnt.load(),
                    m_type_1_largest_time.load());
        }
        if (m_type_2_cnt > 0) {
            fprintf(stdout, "Reconstruction Count: %ld\tReconstruction Avg. Latency: %ld\tMax Recon. Latency:%ld\n",
                    m_type_2_cnt.load(),
                    m_type_2_total_time.load() / m_type_2_cnt.load(),
                    m_type_2_largest_time.load());
        }
    }

private:
    std::mutex m_mutex;
    std::unordered_map<size_t, JobInfo> m_jobs;
    std::vector<Event> m_event_log;

    std::atomic<size_t> m_type_1_cnt;
    std::atomic<size_t> m_type_1_total_time;

    std::atomic<size_t> m_type_2_cnt;
    std::atomic<size_t> m_type_2_total_time;

    std::atomic<size_t> m_type_1_largest_time;
    std::atomic<size_t> m_type_2_largest_time;
};
}