mirror of https://gitee.com/anolis/sysom.git
refactor(sysom): Remove sysom_node_init config
This commit is contained in:
parent
f534733d25
commit
275887648f
|
@ -154,23 +154,6 @@ download_node_exporter()
|
|||
|
||||
}
|
||||
|
||||
prepare_node_init_tar()
|
||||
{
|
||||
mkdir -p ${NODE_INIT_DIR}
|
||||
cp -r ${BASE_DIR}/../../node/monitor ${NODE_INIT_DIR}
|
||||
cp ${RESOURCE_DIR}/${NODE_EXPORTER_TAR} ${NODE_INIT_DIR}/monitor/
|
||||
}
|
||||
|
||||
set_node_init_cmd()
|
||||
{
|
||||
sed "s#server_local_ip='xxx'#server_local_ip=\"${SERVER_LOCAL_IP}\"#g" -i ${NODE_INIT_SCRIPT}
|
||||
sed "s#server_public_ip='xxx'#server_public_ip=\"${SERVER_PUBLIC_IP}\"#g" -i ${NODE_INIT_SCRIPT}
|
||||
sed "s#server_port='xxx'#server_port=\"${SERVER_PORT}\"#g" -i ${NODE_INIT_SCRIPT}
|
||||
sed "s#app_home='xxx'#app_home=\"${APP_HOME}\"#g" -i ${NODE_INIT_SCRIPT}
|
||||
sed "s#node_home='xxx'#node_home=\"${NODE_HOME}\"#g" -i ${NODE_DELETE_SCRIPT}
|
||||
}
|
||||
|
||||
|
||||
configure_grafana()
|
||||
{
|
||||
bash -x grafana_api_set.sh
|
||||
|
@ -199,9 +182,6 @@ main()
|
|||
start_grafana_service
|
||||
start_prometheus_service
|
||||
|
||||
set_node_init_cmd
|
||||
#prepare_node_init_tar
|
||||
|
||||
configure_grafana
|
||||
configure_cron
|
||||
|
||||
|
|
|
@ -60,14 +60,7 @@ start_cron()
|
|||
echo "* * * * * pushd ${SERVER_HOME}/vmcore;${VIRTUALENV_PYTHON3} parse_panic.py ${file_path} ${SERVER_PORT};popd" >> /var/spool/cron/root
|
||||
}
|
||||
|
||||
prepare_node_init_tar()
|
||||
{
|
||||
mkdir -p ${NODE_INIT_DIR}
|
||||
cp -r ${BASE_DIR}/../../node/vmcore ${NODE_INIT_DIR}
|
||||
}
|
||||
|
||||
deploy() {
|
||||
# prepare_node_init_tar
|
||||
source_virtualenv
|
||||
init_conf
|
||||
start_app
|
||||
|
|
|
@ -40,22 +40,7 @@ start_app() {
|
|||
exit 1
|
||||
}
|
||||
|
||||
prepare_node_init_tar()
|
||||
{
|
||||
mkdir -p ${NODE_INIT_DIR}
|
||||
cp -r ${BASE_DIR}/../../node/diagnosis ${NODE_INIT_DIR}
|
||||
# pushd ${NODE_INIT_DIR}/diagnosis
|
||||
# ls | grep ${SYSAK_PKG} 1>/dev/null 2>/dev/null
|
||||
# if [ $? -ne 0 ]
|
||||
# then
|
||||
# wget ${SYSAK_DOWNLOAD_URL}/${SYSAK_PKG}
|
||||
# ls
|
||||
# fi
|
||||
# popd
|
||||
}
|
||||
|
||||
deploy() {
|
||||
# prepare_node_init_tar
|
||||
source_virtualenv
|
||||
init_conf
|
||||
start_app
|
||||
|
|
|
@ -43,22 +43,6 @@ NODE_INIT_DIR=sysom_node_init
|
|||
NODE_INIT_PKG=sysom_node_init.tar.gz
|
||||
NODE_DIR=${basedir}/../node
|
||||
|
||||
###initial download sysom_node_init.tar.gz###
|
||||
init_sysom_node_init()
|
||||
{
|
||||
mkdir -p ${UPLOAD_DIR}/${NODE_INIT_DIR}
|
||||
cp ${NODE_DIR}/init.sh ${UPLOAD_DIR}/${NODE_INIT_DIR}
|
||||
cp ${NODE_DIR}/clear.sh ${UPLOAD_DIR}/${NODE_INIT_DIR}
|
||||
}
|
||||
|
||||
tar_sysom_node_init()
|
||||
{
|
||||
pushd ${UPLOAD_DIR}
|
||||
tar -zcf ${NODE_INIT_PKG} ${NODE_INIT_DIR}
|
||||
rm -rf ${NODE_INIT_DIR}
|
||||
popd
|
||||
}
|
||||
|
||||
###enable the service web menu###
|
||||
setup_web_menu_enable()
|
||||
{
|
||||
|
@ -94,7 +78,6 @@ pushd $basedir
|
|||
if [ $FIRST_INIT_DONE == 0 ]
|
||||
then
|
||||
generate_service_env
|
||||
init_sysom_node_init
|
||||
update_global_config
|
||||
for dir in `cat $config`
|
||||
do
|
||||
|
@ -106,7 +89,6 @@ then
|
|||
popd
|
||||
fi
|
||||
done
|
||||
tar_sysom_node_init
|
||||
sed -i 's/^FIRST_INIT_DONE=0/FIRST_INIT_DONE=1/g' $0
|
||||
else
|
||||
for dir in `ls`
|
||||
|
|
|
@ -5,7 +5,7 @@ from django.conf import settings
|
|||
from cec_base.cec_client import MultiConsumer, CecAsyncConsumeTask
|
||||
from cec_base.event import Event
|
||||
from cec_base.consumer import Consumer
|
||||
from sysom_utils import ConfigParser, CecTarget
|
||||
from sysom_utils import ConfigParser
|
||||
from .helper import DiagnosisHelper
|
||||
|
||||
|
||||
|
@ -17,12 +17,6 @@ class DiagnosisTaskExecutor(MultiConsumer):
|
|||
custom_callback=self.on_receive_event
|
||||
)
|
||||
self._config = config
|
||||
self.append_group_consume_task(
|
||||
config.get_server_config().cec.topics.SYSOM_CEC_PLUGIN_TOPIC,
|
||||
settings.SYSOM_CEC_DIAGNOSIS_CONSUMER_GROUP,
|
||||
Consumer.generate_consumer_id(),
|
||||
ensure_topic_exist=True
|
||||
)
|
||||
self.append_group_consume_task(
|
||||
settings.SYSOM_CEC_DIAGNOSIS_TASK_DISPATCH_TOPIC,
|
||||
settings.SYSOM_CEC_DIAGNOSIS_CONSUMER_GROUP,
|
||||
|
@ -35,8 +29,6 @@ class DiagnosisTaskExecutor(MultiConsumer):
|
|||
try:
|
||||
if task.topic_name == settings.SYSOM_CEC_DIAGNOSIS_TASK_DISPATCH_TOPIC:
|
||||
self._process_task_dispatch_event(event)
|
||||
elif task.topic_name == self._config.get_server_config().cec.topics.SYSOM_CEC_PLUGIN_TOPIC:
|
||||
self._process_plugin_event(event)
|
||||
else:
|
||||
# Unexpected
|
||||
logger.error("Receive unknown topic event, unexpected!!")
|
||||
|
@ -48,46 +40,6 @@ class DiagnosisTaskExecutor(MultiConsumer):
|
|||
################################################################################################
|
||||
# 事件处理
|
||||
################################################################################################
|
||||
def _process_plugin_event(self, event: Event):
|
||||
"""Process plugin event
|
||||
{
|
||||
"type": "clean",
|
||||
"params": {
|
||||
"channel": "ssh",
|
||||
"host": instance.ip,
|
||||
"username": instance.username,
|
||||
"port": instance.port
|
||||
},
|
||||
"echo": {
|
||||
"instance": params.get("host", "Unknown host"),
|
||||
"label": "host_init"
|
||||
}
|
||||
}
|
||||
"""
|
||||
from lib.authentications import decode_token
|
||||
try:
|
||||
value = event.value
|
||||
plugin_event_type = value.get("type", "Unknown type")
|
||||
params = value.get("params", {})
|
||||
token = params.pop("token", "")
|
||||
|
||||
if plugin_event_type == "init":
|
||||
params["service_name"] = "node_init"
|
||||
elif plugin_event_type == "clean":
|
||||
params["service_name"] = "node_delete"
|
||||
else:
|
||||
raise Exception(f"Receive not supprt plugin event: {event}")
|
||||
|
||||
user = decode_token(token)
|
||||
# 1. Perform init
|
||||
instance = DiagnosisHelper.init(params, user)
|
||||
|
||||
# 2. Execute diagnosis task
|
||||
self._execute_diagnosis_task_by_model(instance)
|
||||
except Exception as exc:
|
||||
logger.exception(
|
||||
f"Diagnosis process plugin event error: {str(exc)}")
|
||||
|
||||
def _process_task_dispatch_event(self, event: Event):
|
||||
"""Process diagnosis task dispatch event
|
||||
{
|
||||
|
|
|
@ -1,25 +0,0 @@
|
|||
#!/usr/bin/python3
|
||||
# coding=utf-8
|
||||
import json
|
||||
import sys
|
||||
|
||||
###############################################################################
|
||||
## 主机清理脚本
|
||||
###############################################################################
|
||||
node_home='xxx'
|
||||
arg = json.loads(sys.argv[1])
|
||||
host_ip = arg.get('instance')
|
||||
|
||||
result = {}
|
||||
result['commands'] = []
|
||||
|
||||
cmd0 = {}
|
||||
cmd0['instance'] = host_ip
|
||||
do_clear_cmd = "pushd "+node_home+" && bash -x sysom_node_init/clear.sh"
|
||||
del_dir_cmd = "popd && rm -rf "+node_home
|
||||
cmd0['cmd'] = do_clear_cmd+" && "+del_dir_cmd
|
||||
|
||||
result['commands'].append(cmd0)
|
||||
|
||||
data = json.dumps(result)
|
||||
print(data)
|
|
@ -1,29 +0,0 @@
|
|||
#!/usr/bin/python3
|
||||
# coding=utf-8
|
||||
import sys
|
||||
import json
|
||||
|
||||
|
||||
def node_delete_result(raw):
|
||||
postprocess_result = {
|
||||
"code": 0,
|
||||
"err_msg": "",
|
||||
"result": {}
|
||||
}
|
||||
postprocess_result["result"] = {
|
||||
"CommandResult": {"data": [{"key": "", "value": raw}]}
|
||||
}
|
||||
print(json.dumps(postprocess_result, indent=4))
|
||||
|
||||
|
||||
def extract_params():
|
||||
path, res, task_id = sys.argv[1], "", sys.argv[2]
|
||||
with open(path, 'r') as tmp:
|
||||
for line in tmp.readlines():
|
||||
res = res + "\r" + line
|
||||
return res, task_id
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
res, _ = extract_params()
|
||||
node_delete_result(res)
|
|
@ -1,43 +0,0 @@
|
|||
#!/usr/bin/python3
|
||||
# coding=utf-8
|
||||
import json
|
||||
import sys
|
||||
|
||||
###############################################################################
|
||||
## 主机初始化脚本
|
||||
## 主要完成以下功能
|
||||
## 1、增加配置文件
|
||||
## 2、下载主机初始化文件
|
||||
## 3、执行主机初始化操作
|
||||
###############################################################################
|
||||
server_local_ip='xxx'
|
||||
server_public_ip='xxx'
|
||||
app_home='xxx'
|
||||
server_port='xxx'
|
||||
node_home=app_home+"/node"
|
||||
server_home=app_home+"/server"
|
||||
arg = json.loads(sys.argv[1])
|
||||
host_ip = arg.get('instance')
|
||||
|
||||
result = {}
|
||||
result['commands'] = []
|
||||
|
||||
cmd0 = {}
|
||||
cmd0['instance'] = host_ip
|
||||
mkdir_cmd = "rm -rf "+node_home+" && mkdir -p "+node_home+" && cd "+node_home
|
||||
download_res_cmd = "wget -T 3 -t 1 http://`env | grep SSH_CONNECTION | awk '{print $1}' | awk -F\"=\" '{print $NF}'`:"+server_port+"/download/sysom_node_init.tar.gz"
|
||||
do_check_wget = "if [ ! -e sysom_node_init.tar.gz ];then exit 1; fi && tar -xf sysom_node_init.tar.gz && cd sysom_node_init"
|
||||
do_init_conf = "echo SERVER_LOCAL_IP="+server_local_ip+" > conf"
|
||||
do_init_conf = do_init_conf + " && echo SERVER_PUBLIC_IP="+server_public_ip+" >> conf"
|
||||
do_init_conf = do_init_conf + " && echo SERVER_PORT="+server_port+" >> conf"
|
||||
do_init_conf = do_init_conf + " && echo APP_HOME="+app_home+" >> conf"
|
||||
do_init_conf = do_init_conf + " && echo SERVER_HOME="+server_home+" >> conf"
|
||||
do_init_conf = do_init_conf + " && echo NODE_HOME="+node_home+" >> conf"
|
||||
do_init_conf = do_init_conf + " && echo NODE_IP="+host_ip+" >> conf"
|
||||
do_init_cmd = "bash -x init.sh"
|
||||
cmd0['cmd'] = mkdir_cmd+" && "+download_res_cmd+" && "+do_check_wget+" && "+do_init_conf+" && "+do_init_cmd
|
||||
|
||||
result['commands'].append(cmd0)
|
||||
|
||||
data = json.dumps(result)
|
||||
print(data)
|
|
@ -1,29 +0,0 @@
|
|||
#!/usr/bin/python3
|
||||
# coding=utf-8
|
||||
import sys
|
||||
import json
|
||||
|
||||
|
||||
def node_init_result(raw):
|
||||
postprocess_result = {
|
||||
"code": 0,
|
||||
"err_msg": "",
|
||||
"result": {}
|
||||
}
|
||||
postprocess_result["result"] = {
|
||||
"CommandResult": {"data": [{"key": "", "value": raw}]}
|
||||
}
|
||||
print(json.dumps(postprocess_result, indent=4))
|
||||
|
||||
|
||||
def extract_params():
|
||||
path, res, task_id = sys.argv[1], "", sys.argv[2]
|
||||
with open(path, 'r') as tmp:
|
||||
for line in tmp.readlines():
|
||||
res = res + "\r" + line
|
||||
return res, task_id
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
res, _ = extract_params()
|
||||
node_init_result(res)
|
|
@ -5,12 +5,12 @@ import socket
|
|||
|
||||
nfs_ip = '127.0.0.1'
|
||||
nfs_dir = '/usr/vmcore-nfs'
|
||||
node_conf = '/usr/local/sysom/conf'
|
||||
# node_conf = '/usr/local/sysom/conf'
|
||||
if len(sys.argv) >= 3 :
|
||||
nfs_ip = sys.argv[1]
|
||||
nfs_dir = sys.argv[2]
|
||||
if len(sys.argv) == 4 and sys.argv[3] != "":
|
||||
node_conf = sys.argv[3]
|
||||
# if len(sys.argv) == 4 and sys.argv[3] != "":
|
||||
# node_conf = sys.argv[3]
|
||||
|
||||
def get_crash_path():
|
||||
try:
|
||||
|
@ -102,18 +102,19 @@ def upload_nfs(vmcore_dir):
|
|||
pass
|
||||
|
||||
def nfs_config():
|
||||
global nfs_ip, nfs_dir,node_conf
|
||||
global nfs_ip, nfs_dir
|
||||
# global node_conf
|
||||
server_local_ip = ""
|
||||
server_port = "80"
|
||||
try:
|
||||
with open(node_conf,'r') as fin:
|
||||
line = fin.readline()
|
||||
while len(line):
|
||||
if line.startswith("SERVER_LOCAL_IP"):
|
||||
server_local_ip = line.split("SERVER_LOCAL_IP=")[1].strip()
|
||||
if line.startswith("SERVER_PORT"):
|
||||
server_port = line.split("SERVER_PORT=")[1].strip()
|
||||
line = fin.readline()
|
||||
# with open(node_conf,'r') as fin:
|
||||
# line = fin.readline()
|
||||
# while len(line):
|
||||
# if line.startswith("SERVER_LOCAL_IP"):
|
||||
# server_local_ip = line.split("SERVER_LOCAL_IP=")[1].strip()
|
||||
# if line.startswith("SERVER_PORT"):
|
||||
# server_port = line.split("SERVER_PORT=")[1].strip()
|
||||
# line = fin.readline()
|
||||
if server_local_ip != "":
|
||||
cmd = f'wget -T 3 -t 1 http://{server_local_ip}:{server_port}/download/vmcore_nfs_config -O vmcore_nfs_config'
|
||||
ret = os.system(cmd)
|
||||
|
|
Loading…
Reference in New Issue