Unverified Commit 7f846576 authored by Eric_Lee's avatar Eric_Lee Committed by GitHub

Merge pull request #133 from jumpserver/dev_bugfix

Dev bugfix
parents 3ed7a99f 3a5cc5a2
...@@ -28,7 +28,7 @@ type Client struct { ...@@ -28,7 +28,7 @@ type Client struct {
BaseHost string BaseHost string
basicAuth []string basicAuth []string
cookie map[string]string cookie map[string]string
http *http.Client http http.Client
UrlParsers []UrlParser UrlParsers []UrlParser
} }
...@@ -36,11 +36,12 @@ type UrlParser interface { ...@@ -36,11 +36,12 @@ type UrlParser interface {
parse(url string, params ...map[string]string) string parse(url string, params ...map[string]string) string
} }
func NewClient(timeout time.Duration, baseHost string) *Client { func NewClient(timeout time.Duration, baseHost string) Client {
headers := make(map[string]string, 0) headers := make(map[string]string, 0)
client := new(http.Client) client := http.Client{
client.Timeout = timeout * time.Second Timeout: timeout * time.Second,
return &Client{ }
return Client{
BaseHost: baseHost, BaseHost: baseHost,
Timeout: timeout * time.Second, Timeout: timeout * time.Second,
Headers: headers, Headers: headers,
......
...@@ -148,14 +148,14 @@ func (t *WrapperTable) convertDataToSlice() [][]string { ...@@ -148,14 +148,14 @@ func (t *WrapperTable) convertDataToSlice() [][]string {
switch t.TruncPolicy { switch t.TruncPolicy {
case TruncSuffix: case TruncSuffix:
row[m] = fmt.Sprintf("%s...", row[m] = fmt.Sprintf("%s...",
GetValidString(j[n],columSize-3, true)) GetValidString(j[n], columSize-3, true))
case TruncPrefix: case TruncPrefix:
row[m] = fmt.Sprintf("...%s", row[m] = fmt.Sprintf("...%s",
GetValidString(j[n],len(j[n])-columSize-3, false)) GetValidString(j[n], len(j[n])-columSize-3, false))
case TruncMiddle: case TruncMiddle:
midValue := (columSize - 3) / 2 midValue := (columSize - 3) / 2
row[m] = fmt.Sprintf("%s...%s",GetValidString(j[n],midValue, true), row[m] = fmt.Sprintf("%s...%s", GetValidString(j[n], midValue, true),
GetValidString(j[n],len(j[n])-midValue, false)) GetValidString(j[n], len(j[n])-midValue, false))
} }
} }
...@@ -168,8 +168,8 @@ func (t *WrapperTable) convertDataToSlice() [][]string { ...@@ -168,8 +168,8 @@ func (t *WrapperTable) convertDataToSlice() [][]string {
func (t *WrapperTable) Display() string { func (t *WrapperTable) Display() string {
t.CalculateColumnsSize() t.CalculateColumnsSize()
tableString := &strings.Builder{} tableString := strings.Builder{}
table := tablewriter.NewWriter(tableString) table := tablewriter.NewWriter(&tableString)
table.SetBorder(false) table.SetBorder(false)
table.SetHeader(t.Labels) table.SetHeader(t.Labels)
colors := make([]tablewriter.Colors, len(t.Fields)) colors := make([]tablewriter.Colors, len(t.Fields))
...@@ -193,23 +193,23 @@ func (t *WrapperTable) Display() string { ...@@ -193,23 +193,23 @@ func (t *WrapperTable) Display() string {
return tableString.String() return tableString.String()
} }
func GetValidString(s string, position int, positive bool) string{ func GetValidString(s string, position int, positive bool) string {
step := 1 step := 1
if positive { if positive {
step = -1 step = -1
} }
for position >=0 && position <= len(s) { for position >= 0 && position <= len(s) {
switch positive { switch positive {
case true: case true:
if utf8.ValidString(s[:position]){ if utf8.ValidString(s[:position]) {
return s[:position] return s[:position]
} }
case false: case false:
if utf8.ValidString(s[position:]){ if utf8.ValidString(s[position:]) {
return s[position:] return s[position:]
} }
} }
position += step position += step
} }
return "" return ""
} }
\ No newline at end of file
...@@ -28,10 +28,12 @@ func GzipCompressFile(srcPath, dstPath string) error { ...@@ -28,10 +28,12 @@ func GzipCompressFile(srcPath, dstPath string) error {
if err != nil { if err != nil {
return err return err
} }
defer sf.Close()
df, err := os.Create(dstPath) df, err := os.Create(dstPath)
if err != nil { if err != nil {
return err return err
} }
defer df.Close()
writer := gzip.NewWriter(df) writer := gzip.NewWriter(df)
writer.Name = dstPath writer.Name = dstPath
writer.ModTime = time.Now().UTC() writer.ModTime = time.Now().UTC()
...@@ -60,6 +62,6 @@ func Abs(x int) int { ...@@ -60,6 +62,6 @@ func Abs(x int) int {
return x return x
} }
func CurrentUTCTime()string{ func CurrentUTCTime() string {
return time.Now().UTC().Format("2006-01-02 15:04:05 +0000") return time.Now().UTC().Format("2006-01-02 15:04:05 +0000")
} }
\ No newline at end of file
...@@ -172,18 +172,14 @@ var Conf = &Config{ ...@@ -172,18 +172,14 @@ var Conf = &Config{
ZipTmpPath: "/tmp", ZipTmpPath: "/tmp",
} }
func SetConf(conf *Config) { func SetConf(conf Config) {
lock.Lock() lock.Lock()
defer lock.Unlock() defer lock.Unlock()
Conf = conf Conf = &conf
} }
func GetConf() *Config { func GetConf() Config {
lock.RLock() lock.RLock()
defer lock.RUnlock() defer lock.RUnlock()
var conf Config return *Conf
if confBytes, err := json.Marshal(Conf); err == nil {
_ = json.Unmarshal(confBytes, &conf)
}
return &conf
} }
...@@ -13,8 +13,8 @@ import ( ...@@ -13,8 +13,8 @@ import (
"github.com/jumpserver/koko/pkg/utils" "github.com/jumpserver/koko/pkg/utils"
) )
func NewAssetPagination(term *utils.Terminal, assets []model.Asset) *AssetPagination { func NewAssetPagination(term *utils.Terminal, assets []model.Asset) AssetPagination {
assetPage := &AssetPagination{term: term, assets: assets} assetPage := AssetPagination{term: term, assets: assets}
assetPage.Initial() assetPage.Initial()
return assetPage return assetPage
} }
...@@ -181,8 +181,8 @@ func (p *AssetPagination) displayTipsInfo() { ...@@ -181,8 +181,8 @@ func (p *AssetPagination) displayTipsInfo() {
} }
func NewUserPagination(term *utils.Terminal, uid, search string, policy bool) *UserAssetPagination { func NewUserPagination(term *utils.Terminal, uid, search string, policy bool) UserAssetPagination {
return &UserAssetPagination{ return UserAssetPagination{
UserID: uid, UserID: uid,
offset: 0, offset: 0,
limit: 0, limit: 0,
...@@ -269,7 +269,7 @@ func (p *UserAssetPagination) Start() []model.Asset { ...@@ -269,7 +269,7 @@ func (p *UserAssetPagination) Start() []model.Asset {
func (p *UserAssetPagination) displayPageAssets() { func (p *UserAssetPagination) displayPageAssets() {
if len(p.Data.Data) == 0 { if len(p.Data.Data) == 0 {
_, _ = p.term.Write([]byte(getI18nFromMap("NoAssets")+"\n\r")) _, _ = p.term.Write([]byte(getI18nFromMap("NoAssets") + "\n\r"))
return return
} }
......
...@@ -43,6 +43,7 @@ func (w *WrapperSession) readLoop() { ...@@ -43,6 +43,7 @@ func (w *WrapperSession) readLoop() {
} }
w.mux.RLock() w.mux.RLock()
_ = w.inWriter.Close() _ = w.inWriter.Close()
_ = w.outReader.Close()
w.mux.RUnlock() w.mux.RUnlock()
close(w.closed) close(w.closed)
logger.Infof("Request %s: Read loop break", w.Uuid) logger.Infof("Request %s: Read loop break", w.Uuid)
......
...@@ -31,8 +31,8 @@ const ( ...@@ -31,8 +31,8 @@ const (
CommandOutputParserName = "Command Output parser" CommandOutputParserName = "Command Output parser"
) )
func newParser(sid string) *Parser { func newParser(sid string) Parser {
parser := &Parser{id: sid} parser := Parser{id: sid}
parser.initial() parser.initial()
return parser return parser
} }
...@@ -55,8 +55,8 @@ type Parser struct { ...@@ -55,8 +55,8 @@ type Parser struct {
command string command string
output string output string
cmdInputParser *CmdParser cmdInputParser CmdParser
cmdOutputParser *CmdParser cmdOutputParser CmdParser
cmdFilterRules []model.SystemUserFilterRule cmdFilterRules []model.SystemUserFilterRule
closed chan struct{} closed chan struct{}
......
...@@ -12,8 +12,8 @@ import ( ...@@ -12,8 +12,8 @@ import (
var ps1Pattern = regexp.MustCompile(`^\[?.*@.*\]?[\\$#]\s|mysql>\s`) var ps1Pattern = regexp.MustCompile(`^\[?.*@.*\]?[\\$#]\s|mysql>\s`)
func NewCmdParser(sid, name string) *CmdParser { func NewCmdParser(sid, name string) CmdParser {
parser := &CmdParser{id: sid, name:name} parser := CmdParser{id: sid, name:name}
parser.initial() parser.initial()
return parser return parser
} }
...@@ -67,16 +67,16 @@ func (cp *CmdParser) initial() { ...@@ -67,16 +67,16 @@ func (cp *CmdParser) initial() {
go func() { go func() {
logger.Infof("Session %s: %s start", cp.id, cp.name) logger.Infof("Session %s: %s start", cp.id, cp.name)
defer logger.Infof("Session %s: %s parser close", cp.id, cp.name) defer logger.Infof("Session %s: %s parser close", cp.id, cp.name)
inloop: loop:
for { for {
line, err := cp.term.ReadLine() line, err := cp.term.ReadLine()
if err != nil { if err != nil {
select { select {
case <-cp.closed: case <-cp.closed:
goto outloop break loop
default: default:
} }
goto inloop goto loop
} }
cp.lock.Lock() cp.lock.Lock()
cp.currentLength += len(line) cp.currentLength += len(line)
...@@ -85,7 +85,6 @@ func (cp *CmdParser) initial() { ...@@ -85,7 +85,6 @@ func (cp *CmdParser) initial() {
} }
cp.lock.Unlock() cp.lock.Unlock()
} }
outloop:
}() }()
} }
......
...@@ -21,8 +21,8 @@ func NewCommandRecorder(sid string) (recorder *CommandRecorder) { ...@@ -21,8 +21,8 @@ func NewCommandRecorder(sid string) (recorder *CommandRecorder) {
return recorder return recorder
} }
func NewReplyRecord(sid string) (recorder *ReplyRecorder) { func NewReplyRecord(sid string) (recorder ReplyRecorder) {
recorder = &ReplyRecorder{SessionID: sid} recorder = ReplyRecorder{SessionID: sid}
recorder.initial() recorder.initial()
return recorder return recorder
} }
......
...@@ -18,7 +18,7 @@ type AzureReplayStorage struct { ...@@ -18,7 +18,7 @@ type AzureReplayStorage struct {
EndpointSuffix string EndpointSuffix string
} }
func (a *AzureReplayStorage) Upload(gZipFilePath, target string) (err error) { func (a AzureReplayStorage) Upload(gZipFilePath, target string) (err error) {
file, err := os.Open(gZipFilePath) file, err := os.Open(gZipFilePath)
if err != nil { if err != nil {
return return
......
...@@ -17,7 +17,7 @@ type ESCommandStorage struct { ...@@ -17,7 +17,7 @@ type ESCommandStorage struct {
DocType string DocType string
} }
func (es *ESCommandStorage) BulkSave(commands []*model.Command) (err error) { func (es ESCommandStorage) BulkSave(commands []*model.Command) (err error) {
var buf bytes.Buffer var buf bytes.Buffer
esClinet, err := elasticsearch.NewClient(elasticsearch.Config{ esClinet, err := elasticsearch.NewClient(elasticsearch.Config{
Addresses: es.Hosts, Addresses: es.Hosts,
......
...@@ -13,7 +13,7 @@ type OSSReplayStorage struct { ...@@ -13,7 +13,7 @@ type OSSReplayStorage struct {
SecretKey string SecretKey string
} }
func (o *OSSReplayStorage) Upload(gZipFilePath, target string) (err error) { func (o OSSReplayStorage) Upload(gZipFilePath, target string) (err error) {
client, err := oss.New(o.Endpoint, o.AccessKey, o.SecretKey) client, err := oss.New(o.Endpoint, o.AccessKey, o.SecretKey)
if err != nil { if err != nil {
return return
......
...@@ -19,7 +19,7 @@ type S3ReplayStorage struct { ...@@ -19,7 +19,7 @@ type S3ReplayStorage struct {
Endpoint string Endpoint string
} }
func (s *S3ReplayStorage) Upload(gZipFilePath, target string) (err error) { func (s S3ReplayStorage) Upload(gZipFilePath, target string) (err error) {
file, err := os.Open(gZipFilePath) file, err := os.Open(gZipFilePath)
if err != nil { if err != nil {
......
...@@ -11,7 +11,7 @@ import ( ...@@ -11,7 +11,7 @@ import (
type ServerCommandStorage struct { type ServerCommandStorage struct {
} }
func (s *ServerCommandStorage) BulkSave(commands []*model.Command) (err error) { func (s ServerCommandStorage) BulkSave(commands []*model.Command) (err error) {
return service.PushSessionCommand(commands) return service.PushSessionCommand(commands)
} }
...@@ -19,7 +19,7 @@ type ServerReplayStorage struct { ...@@ -19,7 +19,7 @@ type ServerReplayStorage struct {
StorageType string StorageType string
} }
func (s *ServerReplayStorage) Upload(gZipFilePath, target string) (err error) { func (s ServerReplayStorage) Upload(gZipFilePath, target string) (err error) {
sessionID := strings.Split(filepath.Base(gZipFilePath), ".")[0] sessionID := strings.Split(filepath.Base(gZipFilePath), ".")[0]
return service.PushSessionReplay(sessionID, gZipFilePath) return service.PushSessionReplay(sessionID, gZipFilePath)
} }
...@@ -117,8 +117,8 @@ func (s *SwitchSession) SetFilterRules(cmdRules []model.SystemUserFilterRule) { ...@@ -117,8 +117,8 @@ func (s *SwitchSession) SetFilterRules(cmdRules []model.SystemUserFilterRule) {
// Bridge 桥接两个链接 // Bridge 桥接两个链接
func (s *SwitchSession) Bridge(userConn UserConnection, srvConn srvconn.ServerConnection) (err error) { func (s *SwitchSession) Bridge(userConn UserConnection, srvConn srvconn.ServerConnection) (err error) {
var ( var (
parser *Parser parser Parser
replayRecorder *ReplyRecorder replayRecorder ReplyRecorder
userInChan chan []byte userInChan chan []byte
srvInChan chan []byte srvInChan chan []byte
......
...@@ -16,8 +16,8 @@ type CommandStorage interface { ...@@ -16,8 +16,8 @@ type CommandStorage interface {
BulkSave(commands []*model.Command) error BulkSave(commands []*model.Command) error
} }
var defaultCommandStorage = &storage.ServerCommandStorage{} var defaultCommandStorage = storage.ServerCommandStorage{}
var defaultReplayStorage = &storage.ServerReplayStorage{StorageType: "server"} var defaultReplayStorage = storage.ServerReplayStorage{StorageType: "server"}
func NewReplayStorage() ReplayStorage { func NewReplayStorage() ReplayStorage {
cf := config.GetConf().ReplayStorage cf := config.GetConf().ReplayStorage
...@@ -31,14 +31,14 @@ func NewReplayStorage() ReplayStorage { ...@@ -31,14 +31,14 @@ func NewReplayStorage() ReplayStorage {
if endpointSuffix == "" { if endpointSuffix == "" {
endpointSuffix = "core.chinacloudapi.cn" endpointSuffix = "core.chinacloudapi.cn"
} }
return &storage.AzureReplayStorage{ return storage.AzureReplayStorage{
AccountName: cf["ACCOUNT_NAME"].(string), AccountName: cf["ACCOUNT_NAME"].(string),
AccountKey: cf["ACCOUNT_KEY"].(string), AccountKey: cf["ACCOUNT_KEY"].(string),
ContainerName: cf["CONTAINER_NAME"].(string), ContainerName: cf["CONTAINER_NAME"].(string),
EndpointSuffix: endpointSuffix, EndpointSuffix: endpointSuffix,
} }
case "oss": case "oss":
return &storage.OSSReplayStorage{ return storage.OSSReplayStorage{
Endpoint: cf["ENDPOINT"].(string), Endpoint: cf["ENDPOINT"].(string),
Bucket: cf["BUCKET"].(string), Bucket: cf["BUCKET"].(string),
AccessKey: cf["ACCESS_KEY"].(string), AccessKey: cf["ACCESS_KEY"].(string),
...@@ -58,7 +58,7 @@ func NewReplayStorage() ReplayStorage { ...@@ -58,7 +58,7 @@ func NewReplayStorage() ReplayStorage {
region = strings.Split(endpoint, ".")[1] region = strings.Split(endpoint, ".")[1]
} }
return &storage.S3ReplayStorage{ return storage.S3ReplayStorage{
Bucket: bucket, Bucket: bucket,
Region: region, Region: region,
AccessKey: cf["ACCESS_KEY"].(string), AccessKey: cf["ACCESS_KEY"].(string),
...@@ -90,7 +90,7 @@ func NewCommandStorage() CommandStorage { ...@@ -90,7 +90,7 @@ func NewCommandStorage() CommandStorage {
if docType == "" { if docType == "" {
docType = "command_store" docType = "command_store"
} }
return &storage.ESCommandStorage{Hosts: hosts, Index: index, DocType: docType} return storage.ESCommandStorage{Hosts: hosts, Index: index, DocType: docType}
default: default:
return defaultCommandStorage return defaultCommandStorage
} }
......
...@@ -35,10 +35,9 @@ func Initial(ctx context.Context) { ...@@ -35,10 +35,9 @@ func Initial(ctx context.Context) {
go KeepSyncConfigWithServer(ctx) go KeepSyncConfigWithServer(ctx)
} }
func newClient() *common.Client { func newClient() common.Client {
cf := config.GetConf() cf := config.GetConf()
cli := common.NewClient(30, "") cli := common.NewClient(30, cf.CoreHost)
cli.BaseHost = cf.CoreHost
return cli return cli
} }
...@@ -87,7 +86,7 @@ func MustLoadServerConfigOnce() { ...@@ -87,7 +86,7 @@ func MustLoadServerConfigOnce() {
func LoadConfigFromServer() (err error) { func LoadConfigFromServer() (err error) {
conf := config.GetConf() conf := config.GetConf()
_, err = authClient.Get(TerminalConfigURL, conf) _, err = authClient.Get(TerminalConfigURL, &conf)
if err != nil { if err != nil {
return err return err
} }
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment