1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
|
package handler
import (
"encoding/json"
"fmt"
"github.com/gorilla/mux"
"io/ioutil"
"net/http"
"strings"
pipeline "hdfs-writer/pkg/pipeline"
utils "hdfs-writer/pkg/utils"
)
var slogger = utils.GetLoggerInstance()
// CreateRouter returns a http handler for the registered URLs
func CreateRouter() http.Handler {
router := mux.NewRouter().StrictSlash(true)
slogger.Info("Created router ...")
router.HandleFunc("/v1/writer", createWriter).Methods("POST")
router.HandleFunc("/v1/writer/{writerName}", deleteWriter).Methods("DELETE")
router.HandleFunc("/v1/writers", getAllWriters).Methods("GET")
return router
}
// CreateWriter creates a pipeline
func createWriter(w http.ResponseWriter, r *http.Request) {
if r.Body == nil {
http.Error(w, "Empty body", http.StatusBadRequest)
return
}
reqBody, err := ioutil.ReadAll(r.Body)
if err != nil {
http.Error(w, err.Error(), http.StatusUnprocessableEntity)
return
}
slogger.Info(string(reqBody))
var results utils.Pipeline
error := json.Unmarshal(reqBody, &results)
if error != nil {
unableToParse := fmt.Sprintf("Could not unmarshal the JSON in create request :: %s", err.Error())
fmt.Fprintln(w, unableToParse)
return
}
if validateKafkaConfig(results.KafkaConfiguration) == false {
http.Error(w, "Validation failed for kafka config items, check logs ..", http.StatusBadRequest)
return
}
if validateHdfsConfig(results.HdfsConfiguration) == false {
http.Error(w, "Validation failed for hdfs config items, check logs ..", http.StatusBadRequest)
return
}
writerName := pipeline.CreatePipeline(results.KafkaConfiguration, results.HdfsConfiguration)
successMessage := fmt.Sprintf("Created the writer ::%s", writerName)
w.Header().Set("Content-Type", "application/json")
w.WriteHeader(http.StatusCreated)
fmt.Fprintln(w, successMessage)
}
// deleteWriter deletes a given writer pipeline
func deleteWriter(w http.ResponseWriter, r *http.Request) {
vars := mux.Vars(r)
writerName := vars["writerName"]
if _, keyExists := pipeline.ChannelMap[writerName]; keyExists {
pipeline.DeletePipeline(writerName)
w.WriteHeader(http.StatusOK)
deleteMessage := fmt.Sprintf("Deleted writer :: %s", writerName)
fmt.Fprintln(w, deleteMessage)
} else {
notFoundMessage := fmt.Sprintf("Could not find writer :: %s", writerName)
fmt.Fprintln(w, notFoundMessage)
}
}
// validateKafkaConfig validates the kafka config items and returns true if they are valid.
func validateKafkaConfig(k utils.KafkaConfig) bool {
if strings.TrimSpace(k.Broker) == "" {
fmt.Println("Broker is empty!")
slogger.Infof("Broker is empty!")
return false
}
if strings.TrimSpace(k.Group) == "" {
fmt.Println("Group is empty!")
slogger.Infof("Group is empty!")
return false
}
if strings.TrimSpace(k.Topic) == "" {
fmt.Println("Topic is empty!")
slogger.Infof("Topic is empty!")
return false
}
return true
}
// validateHdfsConfig validates the kafka config items and returns true if they are valid.
func validateHdfsConfig(h utils.HdfsConfig) bool {
if strings.TrimSpace(h.HdfsURL) == "" {
fmt.Println("HdfsURL is empty!")
return false
}
return true
}
// getAllWriters list down the active writers
func getAllWriters(w http.ResponseWriter, r *http.Request) {
slogger.Info("Listing all the writers ...")
var listOfWriters []string
for k := range pipeline.ChannelMap {
listOfWriters = append(listOfWriters, k)
}
w.Header().Set("Content-Type", "application/json")
w.WriteHeader(http.StatusOK)
w.Write([]byte(fmt.Sprintf(`{"Writers" : "%v"}`, listOfWriters)))
}
|