| @@ -473,3 +473,7 @@ func GetAttachmentSizeByDatasetID(datasetID int64) (int64, error) { | |||
| return total, nil | |||
| } | |||
| func GetAllAttachmentSize() (int64, error) { | |||
| return x.SumInt(&Attachment{}, "size") | |||
| } | |||
| @@ -151,23 +151,42 @@ type TaskPod struct { | |||
| TaskRoleStatus struct { | |||
| Name string `json:"name"` | |||
| } `json:"taskRoleStatus"` | |||
| TaskStatuses []struct { | |||
| TaskIndex int `json:"taskIndex"` | |||
| PodUID string `json:"podUid"` | |||
| PodIP string `json:"podIp"` | |||
| PodName string `json:"podName"` | |||
| ContainerID string `json:"containerId"` | |||
| ContainerIP string `json:"containerIp"` | |||
| ContainerGpus string `json:"containerGpus"` | |||
| State string `json:"state"` | |||
| StartAt time.Time `json:"startAt"` | |||
| FinishedAt time.Time `json:"finishedAt"` | |||
| ExitCode int `json:"exitCode"` | |||
| ExitDiagnostics string `json:"exitDiagnostics"` | |||
| RetriedCount int `json:"retriedCount"` | |||
| StartTime string | |||
| FinishedTime string | |||
| } `json:"taskStatuses"` | |||
| //TaskStatuses []struct { | |||
| // TaskIndex int `json:"taskIndex"` | |||
| // PodUID string `json:"podUid"` | |||
| // PodIP string `json:"podIp"` | |||
| // PodName string `json:"podName"` | |||
| // ContainerID string `json:"containerId"` | |||
| // ContainerIP string `json:"containerIp"` | |||
| // ContainerGpus string `json:"containerGpus"` | |||
| // State string `json:"state"` | |||
| // StartAt time.Time `json:"startAt"` | |||
| // FinishedAt time.Time `json:"finishedAt"` | |||
| // ExitCode int `json:"exitCode"` | |||
| // ExitDiagnostics string `json:"exitDiagnostics"` | |||
| // RetriedCount int `json:"retriedCount"` | |||
| // StartTime string | |||
| // FinishedTime string | |||
| //} `json:"taskStatuses"` | |||
| TaskStatuses []TaskStatuses `json:"taskStatuses"` | |||
| } | |||
| type TaskStatuses struct { | |||
| TaskIndex int `json:"taskIndex"` | |||
| PodUID string `json:"podUid"` | |||
| PodIP string `json:"podIp"` | |||
| PodName string `json:"podName"` | |||
| ContainerID string `json:"containerId"` | |||
| ContainerIP string `json:"containerIp"` | |||
| ContainerGpus string `json:"containerGpus"` | |||
| State string `json:"state"` | |||
| StartAt time.Time `json:"startAt"` | |||
| FinishedAt time.Time `json:"finishedAt"` | |||
| ExitCode int `json:"exitCode"` | |||
| ExitDiagnostics string `json:"exitDiagnostics"` | |||
| RetriedCount int `json:"retriedCount"` | |||
| StartTime string | |||
| FinishedTime string | |||
| } | |||
| type TaskInfo struct { | |||
| @@ -255,6 +274,11 @@ func ConvertToJobResultPayload(input map[string]interface{}) (JobResultPayload, | |||
| err := json.Unmarshal(data, &jobResultPayload) | |||
| jobResultPayload.JobStatus.StartTime = time.Unix(jobResultPayload.JobStatus.CreatedTime/1000, 0).Format("2006-01-02 15:04:05") | |||
| jobResultPayload.JobStatus.EndTime = time.Unix(jobResultPayload.JobStatus.CompletedTime/1000, 0).Format("2006-01-02 15:04:05") | |||
| if jobResultPayload.JobStatus.State == string(JobWaiting) { | |||
| jobResultPayload.JobStatus.StartTime = "-" | |||
| jobResultPayload.JobStatus.EndTime = "-" | |||
| } | |||
| return jobResultPayload, err | |||
| } | |||
| @@ -674,7 +698,7 @@ func GetCloudbrainByName(jobName string) (*Cloudbrain, error) { | |||
| } | |||
| func CanDelJob(isSigned bool, user *User, job *CloudbrainInfo) bool { | |||
| if !isSigned || job.Status != string(JobStopped) { | |||
| if !isSigned || (job.Status != string(JobStopped) && job.Status != string(JobFailed) && job.Status != string(ModelArtsStartFailed) && job.Status != string(ModelArtsCreateFailed)){ | |||
| return false | |||
| } | |||
| repo, err := GetRepositoryByID(job.RepoID) | |||
| @@ -139,7 +139,14 @@ func SearchDatasetCondition(opts *SearchDatasetOptions) builder.Cond { | |||
| if opts.IncludePublic { | |||
| cond = cond.And(builder.Eq{"dataset.status": DatasetStatusPublic}) | |||
| if opts.OwnerID > 0 { | |||
| cond = cond.Or(builder.Eq{"repository.owner_id": opts.OwnerID}) | |||
| if len(opts.Keyword) == 0 { | |||
| cond = cond.Or(builder.Eq{"repository.owner_id": opts.OwnerID}) | |||
| } else { | |||
| subCon := builder.NewCond() | |||
| subCon = subCon.And(builder.Eq{"repository.owner_id": opts.OwnerID}, builder.Like{"dataset.title", opts.Keyword}) | |||
| cond = cond.Or(subCon) | |||
| } | |||
| } | |||
| } else if opts.OwnerID > 0 { | |||
| cond = cond.And(builder.Eq{"repository.owner_id": opts.OwnerID}) | |||
| @@ -137,6 +137,7 @@ func init() { | |||
| tablesStatistic = append(tablesStatistic, | |||
| new(RepoStatistic), | |||
| new(SummaryStatistic), | |||
| new(UserBusinessAnalysis), | |||
| ) | |||
| @@ -1430,6 +1430,15 @@ func GetAllRepositoriesByFilterCols(columns ...string) ([]*Repository, error) { | |||
| } | |||
| func GetAllRepositoriesCount() (int64, error) { | |||
| repo := new(Repository) | |||
| return x.Count(repo) | |||
| } | |||
| func GetAllRepositoriesSize() (int64, error) { | |||
| return x.SumInt(&Repository{}, "size") | |||
| } | |||
| func updateRepository(e Engine, repo *Repository, visibilityChanged bool) (err error) { | |||
| repo.LowerName = strings.ToLower(repo.Name) | |||
| @@ -1,38 +1,62 @@ | |||
| package models | |||
| import ( | |||
| "code.gitea.io/gitea/modules/timeutil" | |||
| "fmt" | |||
| "time" | |||
| "code.gitea.io/gitea/modules/timeutil" | |||
| ) | |||
| // RepoStatistic statistic info of all repository | |||
| type RepoStatistic struct { | |||
| ID int64 `xorm:"pk autoincr"` | |||
| RepoID int64 `xorm:"unique(s) NOT NULL"` | |||
| Date string `xorm:"unique(s) NOT NULL"` | |||
| NumWatches int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumStars int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumForks int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumDownloads int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumComments int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumVisits int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumClosedIssues int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumVersions int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| //develop months | |||
| NumDevMonths int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| RepoSize int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| DatasetSize int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumModels int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumWikiViews int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumCommits int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumIssues int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumPulls int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| IssueFixedRate float32 `xorm:"NOT NULL"` | |||
| NumContributor int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumKeyContributor int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| CreatedUnix timeutil.TimeStamp `xorm:"INDEX created"` | |||
| UpdatedUnix timeutil.TimeStamp `xorm:"INDEX updated"` | |||
| ID int64 `xorm:"pk autoincr"` | |||
| RepoID int64 `xorm:"unique(s) NOT NULL"` | |||
| Date string `xorm:"unique(s) NOT NULL"` | |||
| NumWatches int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumWatchesAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumStars int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumStarsAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumForks int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumForksAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumDownloads int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumDownloadsAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumComments int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumCommentsAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumVisits int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumClosedIssues int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumClosedIssuesAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumVersions int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumDevMonths int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| RepoSize int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| DatasetSize int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumModels int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumWikiViews int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumCommits int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumCommitsAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumIssues int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumIssuesAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumPulls int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumPullsAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| IssueFixedRate float32 `xorm:"NOT NULL"` | |||
| NumContributor int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumContributorAdded int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumKeyContributor int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumContributorsGrowth int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumCommitsGrowth int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumCommitLinesGrowth int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumIssuesGrowth int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumCommentsGrowth int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| Impact float64 `xorm:"NOT NULL DEFAULT 0"` | |||
| Completeness float64 `xorm:"NOT NULL DEFAULT 0"` | |||
| Liveness float64 `xorm:"NOT NULL DEFAULT 0"` | |||
| ProjectHealth float64 `xorm:"NOT NULL DEFAULT 0"` | |||
| TeamHealth float64 `xorm:"NOT NULL DEFAULT 0"` | |||
| Growth float64 `xorm:"NOT NULL DEFAULT 0"` | |||
| RadarTotal float64 `xorm:"NOT NULL DEFAULT 0"` | |||
| CreatedUnix timeutil.TimeStamp `xorm:"INDEX created"` | |||
| UpdatedUnix timeutil.TimeStamp `xorm:"INDEX updated"` | |||
| } | |||
| func DeleteRepoStatDaily(date string) error { | |||
| @@ -55,6 +79,35 @@ func DeleteRepoStatDaily(date string) error { | |||
| return nil | |||
| } | |||
| func GetRepoStatisticByDate(date string) ([]*RepoStatistic, error) { | |||
| repoStatistics := make([]*RepoStatistic, 0) | |||
| err := xStatistic.Where("date = ?", date).Find(&repoStatistics) | |||
| return repoStatistics, err | |||
| } | |||
| func GetOneRepoStatisticBeforeTime(time time.Time) (*RepoStatistic, error) { | |||
| repoStatistics := make([]*RepoStatistic, 0) | |||
| err := xStatistic.Where("created_unix >= ?", time.Unix()).OrderBy("created_unix").Limit(1).Find(&repoStatistics) | |||
| if err != nil { | |||
| return nil, err | |||
| } else { | |||
| if len(repoStatistics) == 0 { | |||
| return nil, fmt.Errorf("the repo statistic record count is 0") | |||
| } else { | |||
| return repoStatistics[0], nil | |||
| } | |||
| } | |||
| } | |||
| func InsertRepoStat(repoStat *RepoStatistic) (int64, error) { | |||
| return xStatistic.Insert(repoStat) | |||
| } | |||
| func UpdateRepoStat(repoStat *RepoStatistic) error { | |||
| sql := "update repo_statistic set impact=?,completeness=?,liveness=?,project_health=?,team_health=?,growth=?,radar_total=? where repo_id=? and date=?" | |||
| _, err := xStatistic.Exec(sql, repoStat.Impact, repoStat.Completeness, repoStat.Liveness, repoStat.ProjectHealth, repoStat.TeamHealth, repoStat.Growth, repoStat.RadarTotal, repoStat.RepoID, repoStat.Date) | |||
| return err | |||
| } | |||
| @@ -0,0 +1,69 @@ | |||
| package models | |||
| import ( | |||
| "fmt" | |||
| "code.gitea.io/gitea/modules/timeutil" | |||
| ) | |||
| var DomainMap = map[string]int{ | |||
| "大模型": 0, | |||
| "ai开发工具": 1, | |||
| "计算机视觉": 2, | |||
| "自然语言处理": 3, | |||
| "机器学习": 4, | |||
| "神经网络": 5, | |||
| "自动驾驶": 6, | |||
| "机器人": 7, | |||
| "联邦学习": 8, | |||
| "数据挖掘": 9, | |||
| "risc-v开发": 10, | |||
| } | |||
| type SummaryStatistic struct { | |||
| ID int64 `xorm:"pk autoincr"` | |||
| Date string `xorm:"unique(s) NOT NULL"` | |||
| NumUsers int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| RepoSize int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| DatasetSize int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumOrganizations int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumModels int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepos int64 `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoBigModel int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoAI int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoVision int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoNLP int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoML int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoNN int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoAutoDrive int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoRobot int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoLeagueLearn int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoDataMining int `xorm:"NOT NULL DEFAULT 0"` | |||
| NumRepoRISC int `xorm:"NOT NULL DEFAULT 0"` | |||
| CreatedUnix timeutil.TimeStamp `xorm:"INDEX created"` | |||
| UpdatedUnix timeutil.TimeStamp `xorm:"INDEX updated"` | |||
| } | |||
| func DeleteSummaryStatisticDaily(date string) error { | |||
| sess := xStatistic.NewSession() | |||
| defer sess.Close() | |||
| if err := sess.Begin(); err != nil { | |||
| return fmt.Errorf("Begin: %v", err) | |||
| } | |||
| if _, err := sess.Where("date = ?", date).Delete(&SummaryStatistic{}); err != nil { | |||
| return fmt.Errorf("Delete: %v", err) | |||
| } | |||
| if err := sess.Commit(); err != nil { | |||
| sess.Close() | |||
| return fmt.Errorf("Commit: %v", err) | |||
| } | |||
| sess.Close() | |||
| return nil | |||
| } | |||
| func InsertSummaryStatistic(summaryStatistic *SummaryStatistic) (int64, error) { | |||
| return xStatistic.Insert(summaryStatistic) | |||
| } | |||
| @@ -98,6 +98,13 @@ func GetTopicByName(name string) (*Topic, error) { | |||
| return &topic, nil | |||
| } | |||
| func GetAllUsedTopics() ([]*Topic, error) { | |||
| topics := make([]*Topic, 0) | |||
| err := x.Where("repo_count > ?", 0).Find(&topics) | |||
| return topics, err | |||
| } | |||
| // addTopicByNameToRepo adds a topic name to a repo and increments the topic count. | |||
| // Returns topic after the addition | |||
| func addTopicByNameToRepo(e Engine, repoID int64, topicName string) (*Topic, error) { | |||
| @@ -178,7 +185,7 @@ func (opts *FindTopicOptions) toConds() builder.Cond { | |||
| } | |||
| if opts.Keyword != "" { | |||
| cond = cond.And(builder.Like{"topic.name", opts.Keyword}) | |||
| cond = cond.And(builder.Like{"topic.name", strings.ToLower(opts.Keyword)}) | |||
| } | |||
| return cond | |||
| @@ -2071,6 +2071,18 @@ func SyncExternalUsers(ctx context.Context, updateExisting bool) error { | |||
| return nil | |||
| } | |||
| func GetUsersCount() (int64, error) { | |||
| user := new(User) | |||
| return x.Where("type=0").Count(user) | |||
| } | |||
| func GetOrganizationsCount() (int64, error) { | |||
| user := new(User) | |||
| return x.Where("type=1").Count(user) | |||
| } | |||
| func GetBlockChainUnSuccessUsers() ([]*User, error) { | |||
| users := make([]*User, 0, 10) | |||
| err := x.Where("public_key = ''"). | |||
| @@ -71,6 +71,49 @@ type UserBusinessAnalysis struct { | |||
| Name string `xorm:"NOT NULL"` | |||
| } | |||
| func QueryUserStaticData(startTime int64, endTime int64) []*UserBusinessAnalysis { | |||
| log.Info("query startTime =" + fmt.Sprint(startTime) + " endTime=" + fmt.Sprint(endTime)) | |||
| statictisSess := xStatistic.NewSession() | |||
| defer statictisSess.Close() | |||
| statictisSess.Select("*").Table("user_business_analysis").Where(" count_date>=" + fmt.Sprint(startTime) + " and count_date<=" + fmt.Sprint(endTime)).OrderBy("count_date desc") | |||
| userBusinessAnalysisList := make([]*UserBusinessAnalysis, 0) | |||
| statictisSess.Find(&userBusinessAnalysisList) | |||
| resultMap := make(map[int64]*UserBusinessAnalysis) | |||
| log.Info("query result size=" + fmt.Sprint(len(userBusinessAnalysisList))) | |||
| for _, userRecord := range userBusinessAnalysisList { | |||
| if _, ok := resultMap[userRecord.ID]; !ok { | |||
| resultMap[userRecord.ID] = userRecord | |||
| } else { | |||
| resultMap[userRecord.ID].CodeMergeCount += userRecord.CodeMergeCount | |||
| resultMap[userRecord.ID].CommitCount += userRecord.CommitCount | |||
| resultMap[userRecord.ID].IssueCount += userRecord.IssueCount | |||
| resultMap[userRecord.ID].CommentCount += userRecord.CommentCount | |||
| resultMap[userRecord.ID].FocusRepoCount += userRecord.FocusRepoCount | |||
| resultMap[userRecord.ID].StarRepoCount += userRecord.StarRepoCount | |||
| resultMap[userRecord.ID].WatchedCount += userRecord.WatchedCount | |||
| resultMap[userRecord.ID].CommitCodeSize += userRecord.CommitCodeSize | |||
| resultMap[userRecord.ID].CommitDatasetSize += userRecord.CommitDatasetSize | |||
| resultMap[userRecord.ID].CommitModelCount += userRecord.CommitModelCount | |||
| resultMap[userRecord.ID].SolveIssueCount += userRecord.SolveIssueCount | |||
| resultMap[userRecord.ID].EncyclopediasCount += userRecord.EncyclopediasCount | |||
| resultMap[userRecord.ID].CreateRepoCount += userRecord.CreateRepoCount | |||
| resultMap[userRecord.ID].LoginCount += userRecord.LoginCount | |||
| } | |||
| } | |||
| userBusinessAnalysisReturnList := make([]*UserBusinessAnalysis, len(resultMap)) | |||
| index := 0 | |||
| for _, v := range resultMap { | |||
| userBusinessAnalysisReturnList[index] = v | |||
| index += 1 | |||
| } | |||
| log.Info("return size=" + fmt.Sprint(len(userBusinessAnalysisReturnList))) | |||
| return userBusinessAnalysisReturnList | |||
| } | |||
| func CountData(wikiCountMap map[string]int) { | |||
| log.Info("start to count other user info data") | |||
| sess := x.NewSession() | |||
| @@ -92,7 +135,7 @@ func CountData(wikiCountMap map[string]int) { | |||
| CountDate := time.Date(currentTimeNow.Year(), currentTimeNow.Month(), currentTimeNow.Day(), 0, 1, 0, 0, currentTimeNow.Location()) | |||
| CodeMergeCountMap := queryAction(start_unix, end_unix, 11) | |||
| CodeMergeCountMap := queryPullRequest(start_unix, end_unix) | |||
| CommitCountMap := queryAction(start_unix, end_unix, 5) | |||
| IssueCountMap := queryAction(start_unix, end_unix, 10) | |||
| @@ -223,6 +266,28 @@ func querySolveIssue(start_unix int64, end_unix int64) map[int64]int { | |||
| } | |||
| func queryPullRequest(start_unix int64, end_unix int64) map[int64]int { | |||
| sess := x.NewSession() | |||
| defer sess.Close() | |||
| sess.Select("issue.*").Table("issue"). | |||
| Join("inner", "pull_request", "issue.id=pull_request.issue_id"). | |||
| Where("pull_request.merged_unix>=" + fmt.Sprint(start_unix) + " and pull_request.merged_unix<=" + fmt.Sprint(end_unix)) | |||
| issueList := make([]*Issue, 0) | |||
| sess.Find(&issueList) | |||
| resultMap := make(map[int64]int) | |||
| log.Info("query issue(PR) size=" + fmt.Sprint(len(issueList))) | |||
| for _, issueRecord := range issueList { | |||
| if _, ok := resultMap[issueRecord.PosterID]; !ok { | |||
| resultMap[issueRecord.PosterID] = 1 | |||
| } else { | |||
| resultMap[issueRecord.PosterID] += 1 | |||
| } | |||
| } | |||
| return resultMap | |||
| } | |||
| func queryAction(start_unix int64, end_unix int64, actionType int64) map[int64]int { | |||
| sess := x.NewSession() | |||
| defer sess.Close() | |||
| @@ -341,7 +406,7 @@ func queryDatasetSize(start_unix int64, end_unix int64) map[int64]int { | |||
| func queryUserCreateRepo(start_unix int64, end_unix int64) map[int64]int { | |||
| sess := x.NewSession() | |||
| defer sess.Close() | |||
| sess.Select("id,owner_id,name").Table("repository").Where(" created_unix>=" + fmt.Sprint(start_unix) + " and created_unix<=" + fmt.Sprint(end_unix)) | |||
| sess.Select("id,owner_id,name").Table("repository").Where("is_fork=false and created_unix>=" + fmt.Sprint(start_unix) + " and created_unix<=" + fmt.Sprint(end_unix)) | |||
| repoList := make([]*Repository, 0) | |||
| sess.Find(&repoList) | |||
| resultMap := make(map[int64]int) | |||
| @@ -354,7 +419,6 @@ func queryUserCreateRepo(start_unix int64, end_unix int64) map[int64]int { | |||
| } | |||
| } | |||
| return resultMap | |||
| } | |||
| func subMonth(t1, t2 time.Time) (month int) { | |||
| @@ -224,7 +224,7 @@ func SizedAvatarLinkWithDomain(email string, size int) string { | |||
| // FileSize calculates the file size and generate user-friendly string. | |||
| func FileSize(s int64) string { | |||
| return humanize.IBytes(uint64(s)) | |||
| return humanize.Bytes(uint64(s)) | |||
| } | |||
| // PrettyNumber produces a string form of the given number in base 10 with | |||
| @@ -310,9 +310,11 @@ func Contexter() macaron.Handler { | |||
| ctx.Data["SignedUserID"] = ctx.User.ID | |||
| ctx.Data["SignedUserName"] = ctx.User.Name | |||
| ctx.Data["IsAdmin"] = ctx.User.IsAdmin | |||
| c.Data["SignedUserName"] = ctx.User.Name | |||
| } else { | |||
| ctx.Data["SignedUserID"] = int64(0) | |||
| ctx.Data["SignedUserName"] = "" | |||
| c.Data["SignedUserName"] = "" | |||
| } | |||
| // If request sends files, parse them here otherwise the Query() can't be parsed and the CsrfToken will be invalid. | |||
| @@ -174,6 +174,16 @@ func registerHandleRepoStatistic() { | |||
| }) | |||
| } | |||
| func registerHandleSummaryStatistic() { | |||
| RegisterTaskFatal("handle_summary_statistic", &BaseConfig{ | |||
| Enabled: true, | |||
| RunAtStart: false, | |||
| Schedule: "@daily", | |||
| }, func(ctx context.Context, _ *models.User, _ Config) error { | |||
| repo.SummaryStatistic() | |||
| return nil | |||
| }) | |||
| } | |||
| func registerHandleUserStatistic() { | |||
| RegisterTaskFatal("handle_user_statistic", &BaseConfig{ | |||
| Enabled: true, | |||
| @@ -202,4 +212,5 @@ func initBasicTasks() { | |||
| registerHandleRepoStatistic() | |||
| registerHandleUserStatistic() | |||
| registerHandleSummaryStatistic() | |||
| } | |||
| @@ -0,0 +1,83 @@ | |||
| package normalization | |||
| import ( | |||
| "code.gitea.io/gitea/modules/setting" | |||
| ) | |||
| func Normalization(value float64, minValue float64, maxValue float64) float64 { | |||
| min := int64(minValue * 100) | |||
| max := int64(maxValue * 100) | |||
| if min == max { | |||
| return 100.0 | |||
| } else { | |||
| return 100 * (value - minValue) / (maxValue - minValue) | |||
| } | |||
| } | |||
| func GetRadarValue(impactValue float64, completeValue float64, livenessValue float64, projectHealthValue float64, teamHealthValue float64, growthValue float64) float64 { | |||
| return setting.RadarMap.Impact*impactValue + | |||
| setting.RadarMap.Completeness*completeValue + | |||
| setting.RadarMap.Liveness*livenessValue + | |||
| setting.RadarMap.ProjectHealth*projectHealthValue + | |||
| setting.RadarMap.TeamHealth*teamHealthValue + | |||
| setting.RadarMap.Growth*growthValue | |||
| } | |||
| func GetImpactInitValue(watch int64, star int64, fork int64, download int64, comments int64, browser int64) float64 { | |||
| return setting.RadarMap.ImpactWatch*float64(watch) + | |||
| setting.RadarMap.ImpactStar*float64(star) + | |||
| setting.RadarMap.ImpactFork*float64(fork) + | |||
| setting.RadarMap.ImpactCodeDownload*float64(download)*0.001 + | |||
| setting.RadarMap.ImpactComments*float64(comments) + | |||
| setting.RadarMap.ImpactBrowser*float64(browser) | |||
| } | |||
| func GetCompleteInitValue(issuesClosed int64, releases int64, developAge int64, dataset int64, model int64, wiki int64) float64 { | |||
| return setting.RadarMap.CompletenessIssuesClosed*float64(issuesClosed) + | |||
| setting.RadarMap.CompletenessReleases*float64(releases) + | |||
| setting.RadarMap.CompletenessDevelopAge*float64(developAge) + | |||
| setting.RadarMap.CompletenessDataset*float64(dataset/(1024*1024)) + | |||
| setting.RadarMap.CompletenessModel*float64(model) + | |||
| setting.RadarMap.CompletenessWiki*float64(wiki) | |||
| } | |||
| func GetLivenessInitValue(commits int64, issues int64, pr int64, release int64) float64 { | |||
| return setting.RadarMap.LivenessCommit*float64(commits) + | |||
| setting.RadarMap.LivenessIssue*float64(issues) + | |||
| setting.RadarMap.LivenessPR*float64(pr) + | |||
| setting.RadarMap.LivenessRelease*float64(release) | |||
| } | |||
| func GetProjectHealthInitValue(issueClosedRatio float32) float64 { | |||
| return setting.RadarMap.ProjectHealthIssueCompleteRatio * float64(issueClosedRatio) | |||
| } | |||
| func GetTeamHealthInitValue(contributors int64, keyContributors int64, newContributors int64) float64 { | |||
| return setting.RadarMap.TeamHealthContributors*float64(contributors) + | |||
| setting.RadarMap.TeamHealthKeyContributors*float64(keyContributors) + | |||
| setting.RadarMap.TeamHealthContributorsAdded*float64(newContributors) | |||
| } | |||
| func GetRepoGrowthInitValue(codelinesGrowth int64, issueGrowth int64, commitsGrowth int64, newContributors int64, commentsGrowth int64) float64 { | |||
| return setting.RadarMap.GrowthCodeLines*float64(codelinesGrowth) + | |||
| setting.RadarMap.GrowthIssue*float64(issueGrowth) + | |||
| setting.RadarMap.GrowthCommit*float64(commitsGrowth) + | |||
| setting.RadarMap.GrowthContributors*float64(newContributors) + | |||
| setting.RadarMap.GrowthComments*float64(commentsGrowth) | |||
| } | |||
| @@ -495,6 +495,47 @@ var ( | |||
| Index string | |||
| TimeField string | |||
| ElkTimeFormat string | |||
| //nginx proxy | |||
| PROXYURL string | |||
| RadarMap = struct { | |||
| Impact float64 | |||
| ImpactWatch float64 | |||
| ImpactStar float64 | |||
| ImpactFork float64 | |||
| ImpactCodeDownload float64 | |||
| ImpactComments float64 | |||
| ImpactBrowser float64 | |||
| Completeness float64 | |||
| CompletenessIssuesClosed float64 | |||
| CompletenessReleases float64 | |||
| CompletenessDevelopAge float64 | |||
| CompletenessDataset float64 | |||
| CompletenessModel float64 | |||
| CompletenessWiki float64 | |||
| Liveness float64 | |||
| LivenessCommit float64 | |||
| LivenessIssue float64 | |||
| LivenessPR float64 | |||
| LivenessRelease float64 | |||
| ProjectHealth float64 | |||
| ProjectHealthIssueCompleteRatio float64 | |||
| TeamHealth float64 | |||
| TeamHealthContributors float64 | |||
| TeamHealthKeyContributors float64 | |||
| TeamHealthContributorsAdded float64 | |||
| Growth float64 | |||
| GrowthCodeLines float64 | |||
| GrowthIssue float64 | |||
| GrowthContributors float64 | |||
| GrowthCommit float64 | |||
| GrowthComments float64 | |||
| }{} | |||
| ) | |||
| // DateLang transforms standard language locale name to corresponding value in datetime plugin. | |||
| @@ -1206,6 +1247,7 @@ func NewContext() { | |||
| Location = sec.Key("LOCATION").MustString("cn-south-222") | |||
| BasePath = sec.Key("BASE_PATH").MustString("attachment/") | |||
| UserBasePath = sec.Key("BASE_PATH_USER").MustString("users/") | |||
| PROXYURL = sec.Key("PROXY_URL").MustString("") | |||
| sec = Cfg.Section("modelarts") | |||
| ModelArtsHost = sec.Key("ENDPOINT").MustString("112.95.163.80") | |||
| @@ -1227,6 +1269,41 @@ func NewContext() { | |||
| Index = sec.Key("INDEX").MustString("filebeat-7.3.2*") | |||
| TimeField = sec.Key("TIMEFIELD").MustString(" @timestamptest") | |||
| ElkTimeFormat = sec.Key("ELKTIMEFORMAT").MustString("date_time") | |||
| sec = Cfg.Section("radar_map") | |||
| RadarMap.Impact = sec.Key("impact").MustFloat64(0.3) | |||
| RadarMap.ImpactWatch = sec.Key("impact_watch").MustFloat64(0.1) | |||
| RadarMap.ImpactStar = sec.Key("impact_star").MustFloat64(0.3) | |||
| RadarMap.ImpactFork = sec.Key("impact_fork").MustFloat64(0.3) | |||
| RadarMap.ImpactCodeDownload = sec.Key("impact_code_download").MustFloat64(0.2) | |||
| RadarMap.ImpactComments = sec.Key("impact_comments").MustFloat64(0.1) | |||
| RadarMap.ImpactBrowser = sec.Key("impact_browser").MustFloat64(0.1) | |||
| RadarMap.Completeness = sec.Key("completeness").MustFloat64(0.1) | |||
| RadarMap.CompletenessIssuesClosed = sec.Key("completeness_issues_closed").MustFloat64(0.2) | |||
| RadarMap.CompletenessReleases = sec.Key("completeness_releases").MustFloat64(0.3) | |||
| RadarMap.CompletenessDevelopAge = sec.Key("completeness_develop_age").MustFloat64(0.1) | |||
| RadarMap.CompletenessDataset = sec.Key("completeness_dataset").MustFloat64(0.1) | |||
| RadarMap.CompletenessModel = sec.Key("completeness_model").MustFloat64(0.1) | |||
| RadarMap.CompletenessWiki = sec.Key("completeness_wiki").MustFloat64(0.1) | |||
| RadarMap.Liveness = sec.Key("liveness").MustFloat64(0.3) | |||
| RadarMap.LivenessCommit = sec.Key("liveness_commit").MustFloat64(0.2) | |||
| RadarMap.LivenessIssue = sec.Key("liveness_issue").MustFloat64(0.2) | |||
| RadarMap.LivenessPR = sec.Key("liveness_pr").MustFloat64(0.2) | |||
| RadarMap.LivenessRelease = sec.Key("liveness_release").MustFloat64(0.4) | |||
| RadarMap.ProjectHealth = sec.Key("project_health").MustFloat64(0.1) | |||
| RadarMap.ProjectHealthIssueCompleteRatio = sec.Key("project_health_issue_complete_ratio").MustFloat64(100) | |||
| RadarMap.TeamHealth = sec.Key("team_health").MustFloat64(0.1) | |||
| RadarMap.TeamHealthContributors = sec.Key("team_health_contributors").MustFloat64(0.2) | |||
| RadarMap.TeamHealthKeyContributors = sec.Key("team_health_key_contributors").MustFloat64(0.6) | |||
| RadarMap.TeamHealthContributorsAdded = sec.Key("team_health_contributors_added").MustFloat64(0.2) | |||
| RadarMap.Growth = sec.Key("growth").MustFloat64(0.1) | |||
| RadarMap.GrowthCodeLines = sec.Key("growth_code_lines").MustFloat64(0.2) | |||
| RadarMap.GrowthIssue = sec.Key("growth_issue").MustFloat64(0.2) | |||
| RadarMap.GrowthContributors = sec.Key("growth_contributors").MustFloat64(0.2) | |||
| RadarMap.GrowthCommit = sec.Key("growth_commit").MustFloat64(0.2) | |||
| RadarMap.GrowthComments = sec.Key("growth_comments").MustFloat64(0.2) | |||
| } | |||
| func loadInternalToken(sec *ini.Section) string { | |||
| @@ -5,11 +5,14 @@ | |||
| package storage | |||
| import ( | |||
| "github.com/unknwon/com" | |||
| "fmt" | |||
| "io" | |||
| "path" | |||
| "strconv" | |||
| "strings" | |||
| "github.com/unknwon/com" | |||
| "code.gitea.io/gitea/modules/log" | |||
| "code.gitea.io/gitea/modules/obs" | |||
| "code.gitea.io/gitea/modules/setting" | |||
| @@ -102,6 +105,49 @@ func CompleteObsMultiPartUpload(uuid, uploadID, fileName string) error { | |||
| return nil | |||
| } | |||
| func ObsMultiPartUpload(uuid string, uploadId string, partNumber int, fileName string, putBody io.ReadCloser) error { | |||
| input := &obs.UploadPartInput{} | |||
| input.Bucket = setting.Bucket | |||
| input.Key = strings.TrimPrefix(path.Join(setting.BasePath, path.Join(uuid[0:1], uuid[1:2], uuid, fileName)), "/") | |||
| input.UploadId = uploadId | |||
| input.PartNumber = partNumber | |||
| input.Body = putBody | |||
| output, err := ObsCli.UploadPart(input) | |||
| if err == nil { | |||
| log.Info("RequestId:%s\n", output.RequestId) | |||
| log.Info("ETag:%s\n", output.ETag) | |||
| return nil | |||
| } else { | |||
| if obsError, ok := err.(obs.ObsError); ok { | |||
| log.Info(obsError.Code) | |||
| log.Info(obsError.Message) | |||
| return obsError | |||
| } else { | |||
| log.Error("error:", err.Error()) | |||
| return err | |||
| } | |||
| } | |||
| } | |||
| func ObsDownload(uuid string, fileName string) (io.ReadCloser, error) { | |||
| input := &obs.GetObjectInput{} | |||
| input.Bucket = setting.Bucket | |||
| input.Key = strings.TrimPrefix(path.Join(setting.BasePath, path.Join(uuid[0:1], uuid[1:2], uuid, fileName)), "/") | |||
| output, err := ObsCli.GetObject(input) | |||
| if err == nil { | |||
| log.Info("StorageClass:%s, ETag:%s, ContentType:%s, ContentLength:%d, LastModified:%s\n", | |||
| output.StorageClass, output.ETag, output.ContentType, output.ContentLength, output.LastModified) | |||
| return output.Body, nil | |||
| } else if obsError, ok := err.(obs.ObsError); ok { | |||
| fmt.Printf("Code:%s\n", obsError.Code) | |||
| fmt.Printf("Message:%s\n", obsError.Message) | |||
| return nil, obsError | |||
| } else { | |||
| return nil, err | |||
| } | |||
| } | |||
| func ObsGenMultiPartSignedUrl(uuid string, uploadId string, partNumber int, fileName string) (string, error) { | |||
| input := &obs.CreateSignedUrlInput{} | |||
| @@ -776,6 +776,54 @@ cloudbrain_creator=创建者 | |||
| cloudbrain_task=任务名称 | |||
| cloudbrain_operate=操作 | |||
| cloudbrain_status_createtime=状态/创建时间 | |||
| cloudbrain_jobname_err=只能以小写字母或数字开头且只包含小写字母、数字、_和-,不能以_结尾,最长36个字符。 | |||
| modelarts.notebook=调试作业 | |||
| modelarts.train_job=训练作业 | |||
| modelarts.train_job.new=新建作业 | |||
| modelarts.train_job.basic_info=基本信息 | |||
| modelarts.train_job.job_status=作业状态 | |||
| modelarts.train_job.job_name=作业名称 | |||
| modelarts.train_job.version=作业版本 | |||
| modelarts.train_job.start_time=开始时间 | |||
| modelarts.train_job.dura_time=持续时间 | |||
| modelarts.train_job.description=作业描述 | |||
| modelarts.train_job.parameter_setting=参数设置 | |||
| modelarts.train_job.parameter_setting_info=参数信息 | |||
| modelarts.train_job.fast_parameter_setting=一键式参数配置 | |||
| modelarts.train_job.fast_parameter_setting_config=如您已保存过参数配置,可单击 | |||
| modelarts.train_job.fast_parameter_setting_config_link=这里 | |||
| modelarts.train_job.frames=常用框架 | |||
| modelarts.train_job.algorithm_origin=算法来源 | |||
| modelarts.train_job.AI_driver=AI引擎 | |||
| modelarts.train_job.start_file=启动文件 | |||
| modelarts.train_job.boot_file_helper=启动文件是您程序执行的入口文件,必须是以.py结尾的文件。 | |||
| modelarts.train_job.boot_file_place=填写启动文件路径,默认为train.py | |||
| modelarts.train_job.dataset=数据集 | |||
| modelarts.train_job.run_parameter=运行参数 | |||
| modelarts.train_job.add_run_parameter=增加运行参数 | |||
| modelarts.train_job.parameter_name=参数名 | |||
| modelarts.train_job.parameter_value=参数值 | |||
| modelarts.train_job.resource_setting=资源设置 | |||
| modelarts.train_job.resource_setting_info=资源信息 | |||
| modelarts.train_job.resource_pool=资源池 | |||
| modelarts.train_job.resource_type=资源类型 | |||
| modelarts.train_job.standard=规格 | |||
| modelarts.train_job.NAS_address=NAS地址 | |||
| modelarts.train_job.NAS_mount_path=NAS挂载路径 | |||
| modelarts.train_job.query_whether_save_parameter=保存作业参数 | |||
| modelarts.train_job.save_helper=保存当前作业的配置参数,后续您可以使用已保存的配置参数快速创建训练作业。 | |||
| modelarts.train_job.common_frame=常用框架 | |||
| modelarts.train_job.amount_of_compute_node=计算节点个数 | |||
| modelarts.train_job.job_parameter_name=作业参数名称 | |||
| modelarts.train_job.parameter_description=作业参数描述 | |||
| modelarts.log=日志 | |||
| modelarts.version_manage=版本管理 | |||
| modelarts.back=返回 | |||
| modelarts.train_job_para_admin=作业参数管理 | |||
| modelarts.train_job_para.edit=编辑 | |||
| modelarts.train_job_para.connfirm=确定 | |||
| template.items=模板选项 | |||
| template.git_content=Git数据(默认分支) | |||
| @@ -309,11 +309,11 @@ function label_task_create(task_name, relate_task_id, taskType,assign_user_id,la | |||
| success:function(res){ | |||
| console.log(res); | |||
| if(res.code == 0){ | |||
| alert("自动标注任务创建成功!"); | |||
| alert("标注任务创建成功!"); | |||
| createsucced = true; | |||
| } | |||
| else{ | |||
| alert("创建自动标注任务失败," + res.message); | |||
| alert("创建标注任务失败," + res.message); | |||
| createsucced = false; | |||
| } | |||
| }, | |||
| @@ -281,10 +281,10 @@ func ExploreDatasets(ctx *context.Context) { | |||
| } | |||
| pager := context.NewPagination(int(count), opts.PageSize, page, 5) | |||
| ctx.Data["Keyword"] = opts.Keyword | |||
| pager.SetDefaultParams(ctx) | |||
| ctx.Data["Page"] = pager | |||
| ctx.Data["Keyword"] = opts.Keyword | |||
| ctx.Data["Datasets"] = datasets | |||
| ctx.Data["Total"] = count | |||
| ctx.Data["PageIsDatasets"] = true | |||
| @@ -262,10 +262,15 @@ func GetAttachment(ctx *context.Context) { | |||
| return | |||
| } | |||
| } else { | |||
| url, err = storage.ObsGetPreSignedUrl(attach.UUID, attach.Name) | |||
| if err != nil { | |||
| ctx.ServerError("ObsGetPreSignedUrl", err) | |||
| return | |||
| if setting.PROXYURL != "" { | |||
| url = setting.PROXYURL + "/obs_proxy_download?uuid=" + attach.UUID + "&file_name=" + attach.Name | |||
| log.Info("return url=" + url) | |||
| } else { | |||
| url, err = storage.ObsGetPreSignedUrl(attach.UUID, attach.Name) | |||
| if err != nil { | |||
| ctx.ServerError("ObsGetPreSignedUrl", err) | |||
| return | |||
| } | |||
| } | |||
| } | |||
| @@ -273,7 +278,6 @@ func GetAttachment(ctx *context.Context) { | |||
| ctx.ServerError("Update", err) | |||
| return | |||
| } | |||
| http.Redirect(ctx.Resp, ctx.Req.Request, url, http.StatusMovedPermanently) | |||
| } else { | |||
| fr, err := storage.Attachments.Open(attach.RelativePath()) | |||
| @@ -282,7 +286,6 @@ func GetAttachment(ctx *context.Context) { | |||
| return | |||
| } | |||
| defer fr.Close() | |||
| if err = increaseDownloadCount(attach, dataSet); err != nil { | |||
| ctx.ServerError("Update", err) | |||
| return | |||
| @@ -662,6 +665,53 @@ func NewMultipart(ctx *context.Context) { | |||
| } | |||
| } | |||
| func PutOBSProxyUpload(ctx *context.Context) { | |||
| uuid := ctx.Query("uuid") | |||
| uploadID := ctx.Query("uploadId") | |||
| partNumber := ctx.QueryInt("partNumber") | |||
| fileName := ctx.Query("file_name") | |||
| RequestBody := ctx.Req.Body() | |||
| if RequestBody == nil { | |||
| ctx.Error(500, fmt.Sprintf("FormFile: %v", RequestBody)) | |||
| return | |||
| } | |||
| err := storage.ObsMultiPartUpload(uuid, uploadID, partNumber, fileName, RequestBody.ReadCloser()) | |||
| if err != nil { | |||
| log.Info("upload error.") | |||
| } | |||
| } | |||
| func GetOBSProxyDownload(ctx *context.Context) { | |||
| uuid := ctx.Query("uuid") | |||
| fileName := ctx.Query("file_name") | |||
| body, err := storage.ObsDownload(uuid, fileName) | |||
| if err != nil { | |||
| log.Info("upload error.") | |||
| } else { | |||
| defer body.Close() | |||
| ctx.Resp.Header().Set("Content-Disposition", "attachment; filename="+fileName) | |||
| ctx.Resp.Header().Set("Content-Type", "application/octet-stream") | |||
| p := make([]byte, 1024) | |||
| var readErr error | |||
| var readCount int | |||
| // 读取对象内容 | |||
| for { | |||
| readCount, readErr = body.Read(p) | |||
| if readCount > 0 { | |||
| ctx.Resp.Write(p[:readCount]) | |||
| //fmt.Printf("%s", p[:readCount]) | |||
| } | |||
| if readErr != nil { | |||
| break | |||
| } | |||
| } | |||
| } | |||
| } | |||
| func GetMultipartUploadUrl(ctx *context.Context) { | |||
| uuid := ctx.Query("uuid") | |||
| uploadID := ctx.Query("uploadID") | |||
| @@ -689,10 +739,16 @@ func GetMultipartUploadUrl(ctx *context.Context) { | |||
| return | |||
| } | |||
| } else { | |||
| url, err = storage.ObsGenMultiPartSignedUrl(uuid, uploadID, partNumber, fileName) | |||
| if err != nil { | |||
| ctx.Error(500, fmt.Sprintf("ObsGenMultiPartSignedUrl failed: %v", err)) | |||
| return | |||
| if setting.PROXYURL != "" { | |||
| url = setting.PROXYURL + "/obs_proxy_multipart?uuid=" + uuid + "&uploadId=" + uploadID + "&partNumber=" + fmt.Sprint(partNumber) + "&file_name=" + fileName | |||
| log.Info("return url=" + url) | |||
| } else { | |||
| url, err = storage.ObsGenMultiPartSignedUrl(uuid, uploadID, partNumber, fileName) | |||
| if err != nil { | |||
| ctx.Error(500, fmt.Sprintf("ObsGenMultiPartSignedUrl failed: %v", err)) | |||
| return | |||
| } | |||
| log.Info("url=" + url) | |||
| } | |||
| } | |||
| @@ -40,6 +40,8 @@ var ( | |||
| categories *models.Categories | |||
| ) | |||
| var jobNamePattern = regexp.MustCompile(`^[a-z0-9][a-z0-9-_]{1,34}[a-z0-9-]$`) | |||
| // MustEnableDataset check if repository enable internal cb | |||
| func MustEnableCloudbrain(ctx *context.Context) { | |||
| if !ctx.Repo.CanRead(models.UnitTypeCloudBrain) { | |||
| @@ -200,6 +202,11 @@ func CloudBrainCreate(ctx *context.Context, form auth.CreateCloudBrainForm) { | |||
| gpuQueue := setting.JobType | |||
| codePath := setting.JobPath + jobName + cloudbrain.CodeMountPath | |||
| resourceSpecId := form.ResourceSpecId | |||
| if !jobNamePattern.MatchString(jobName) { | |||
| ctx.RenderWithErr(ctx.Tr("repo.cloudbrain_jobname_err"), tplModelArtsNew, &form) | |||
| return | |||
| } | |||
| if jobType != string(models.JobTypeBenchmark) && jobType != string(models.JobTypeDebug) && jobType != string(models.JobTypeSnn4imagenet) && jobType != string(models.JobTypeBrainScore) { | |||
| log.Error("jobtype error:", jobType, ctx.Data["MsgID"]) | |||
| @@ -281,17 +288,30 @@ func CloudBrainShow(ctx *context.Context) { | |||
| if result != nil { | |||
| jobRes, _ := models.ConvertToJobResultPayload(result.Payload) | |||
| jobRes.Resource.Memory = strings.ReplaceAll(jobRes.Resource.Memory, "Mi", "MB") | |||
| ctx.Data["result"] = jobRes | |||
| taskRoles := jobRes.TaskRoles | |||
| taskRes, _ := models.ConvertToTaskPod(taskRoles[cloudbrain.SubTaskName].(map[string]interface{})) | |||
| ctx.Data["taskRes"] = taskRes | |||
| task.Status = taskRes.TaskStatuses[0].State | |||
| task.ContainerID = taskRes.TaskStatuses[0].ContainerID | |||
| task.ContainerIp = taskRes.TaskStatuses[0].ContainerIP | |||
| err = models.UpdateJob(task) | |||
| if err != nil { | |||
| ctx.Data["error"] = err.Error() | |||
| if jobRes.JobStatus.State != string(models.JobFailed) { | |||
| taskRes, _ := models.ConvertToTaskPod(taskRoles[cloudbrain.SubTaskName].(map[string]interface{})) | |||
| ctx.Data["taskRes"] = taskRes | |||
| task.Status = taskRes.TaskStatuses[0].State | |||
| task.ContainerID = taskRes.TaskStatuses[0].ContainerID | |||
| task.ContainerIp = taskRes.TaskStatuses[0].ContainerIP | |||
| err = models.UpdateJob(task) | |||
| if err != nil { | |||
| ctx.Data["error"] = err.Error() | |||
| } | |||
| } else { | |||
| task.Status = jobRes.JobStatus.State | |||
| taskRes := models.TaskPod{TaskStatuses: []models.TaskStatuses{ | |||
| { | |||
| State: jobRes.JobStatus.State, | |||
| }, | |||
| }} | |||
| ctx.Data["taskRes"] = taskRes | |||
| jobRes.JobStatus.StartTime = time.Unix(int64(task.CreatedUnix), 0).Format("2006-01-02 15:04:05") | |||
| jobRes.JobStatus.EndTime = time.Unix(int64(task.UpdatedUnix), 0).Format("2006-01-02 15:04:05") | |||
| } | |||
| ctx.Data["result"] = jobRes | |||
| } | |||
| ctx.Data["task"] = task | |||
| @@ -351,7 +371,7 @@ func CloudBrainStop(ctx *context.Context) { | |||
| return | |||
| } | |||
| if task.Status == string(models.JobStopped) { | |||
| if task.Status == string(models.JobStopped) || task.Status == string(models.JobFailed) { | |||
| log.Error("the job(%s) has been stopped", task.JobName, ctx.Data["msgID"]) | |||
| ctx.ServerError("the job has been stopped", errors.New("the job has been stopped")) | |||
| return | |||
| @@ -454,7 +474,7 @@ func CloudBrainDel(ctx *context.Context) { | |||
| return | |||
| } | |||
| if task.Status != string(models.JobStopped) { | |||
| if task.Status != string(models.JobStopped) && task.Status != string(models.JobFailed){ | |||
| log.Error("the job(%s) has not been stopped", task.JobName, ctx.Data["msgID"]) | |||
| ctx.ServerError("the job has not been stopped", errors.New("the job has not been stopped")) | |||
| return | |||
| @@ -100,7 +100,10 @@ func ModelArtsCreate(ctx *context.Context, form auth.CreateModelArtsForm) { | |||
| uuid := form.Attachment | |||
| description := form.Description | |||
| //repo := ctx.Repo.Repository | |||
| if !jobNamePattern.MatchString(jobName) { | |||
| ctx.RenderWithErr(ctx.Tr("repo.cloudbrain_jobname_err"), tplModelArtsNew, &form) | |||
| return | |||
| } | |||
| err := modelarts.GenerateTask(ctx, jobName, uuid, description) | |||
| if err != nil { | |||
| ctx.RenderWithErr(err.Error(), tplModelArtsNew, &form) | |||
| @@ -228,7 +231,7 @@ func ModelArtsDel(ctx *context.Context) { | |||
| return | |||
| } | |||
| if task.Status != string(models.JobStopped) { | |||
| if task.Status != string(models.ModelArtsCreateFailed) && task.Status != string(models.ModelArtsStartFailed) && task.Status != string(models.ModelArtsStopped){ | |||
| log.Error("the job(%s) has not been stopped", task.JobName) | |||
| ctx.ServerError("the job has not been stopped", errors.New("the job has not been stopped")) | |||
| return | |||
| @@ -3,6 +3,8 @@ package repo | |||
| import ( | |||
| "time" | |||
| "code.gitea.io/gitea/modules/normalization" | |||
| "code.gitea.io/gitea/models" | |||
| "code.gitea.io/gitea/modules/log" | |||
| "code.gitea.io/gitea/modules/repository" | |||
| @@ -17,6 +19,8 @@ func RepoStatisticAuto() { | |||
| func RepoStatisticDaily(date string) { | |||
| log.Info("%s", date) | |||
| log.Info("begin Repo Statistic") | |||
| t, _ := time.Parse("2006-01-02", "date") | |||
| if err := models.DeleteRepoStatDaily(date); err != nil { | |||
| log.Error("DeleteRepoStatDaily failed: %v", err.Error()) | |||
| return | |||
| @@ -28,13 +32,25 @@ func RepoStatisticDaily(date string) { | |||
| return | |||
| } | |||
| for _, repo := range repos { | |||
| var reposRadar = make([]*models.RepoStatistic, 0) | |||
| var minRepoRadar models.RepoStatistic | |||
| var maxRepoRadar models.RepoStatistic | |||
| for i, repo := range repos { | |||
| log.Info("start statistic: %s", repo.Name) | |||
| var numDevMonths, numWikiViews, numContributor, numKeyContributor, numCommitsGrowth, numCommitLinesGrowth, numContributorsGrowth int64 | |||
| repoGitStat, err := models.GetRepoKPIStats(repo) | |||
| if err != nil { | |||
| log.Error("GetRepoKPIStats failed: %s", repo.Name) | |||
| log.Error("failed statistic: %s", repo.Name) | |||
| continue | |||
| } else { | |||
| numDevMonths = repoGitStat.DevelopAge | |||
| numKeyContributor = repoGitStat.KeyContributors | |||
| numWikiViews = repoGitStat.WikiPages | |||
| numContributor = repoGitStat.Contributors | |||
| numCommitsGrowth = repoGitStat.CommitsAdded | |||
| numCommitLinesGrowth = repoGitStat.CommitLinesModified | |||
| numContributorsGrowth = repoGitStat.ContributorsAdded | |||
| } | |||
| var issueFixedRate float32 | |||
| @@ -42,66 +58,178 @@ func RepoStatisticDaily(date string) { | |||
| issueFixedRate = float32(repo.NumClosedIssues) / float32(repo.NumIssues) | |||
| } | |||
| numVersions, err := models.GetReleaseCountByRepoID(repo.ID, models.FindReleasesOptions{}) | |||
| var numVersions int64 | |||
| numVersions, err = models.GetReleaseCountByRepoID(repo.ID, models.FindReleasesOptions{}) | |||
| if err != nil { | |||
| log.Error("GetReleaseCountByRepoID failed: %s", repo.Name) | |||
| log.Error("failed statistic: %s", repo.Name) | |||
| continue | |||
| log.Error("GetReleaseCountByRepoID failed(%s): %v", repo.Name, err) | |||
| } | |||
| datasetSize, err := getDatasetSize(repo) | |||
| var datasetSize int64 | |||
| datasetSize, err = getDatasetSize(repo) | |||
| if err != nil { | |||
| log.Error("getDatasetSize failed: %s", repo.Name) | |||
| log.Error("failed statistic: %s", repo.Name) | |||
| continue | |||
| log.Error("getDatasetSize failed(%s): %v", repo.Name, err) | |||
| } | |||
| numComments, err := models.GetCommentCountByRepoID(repo.ID) | |||
| var numComments int64 | |||
| numComments, err = models.GetCommentCountByRepoID(repo.ID) | |||
| if err != nil { | |||
| log.Error("GetCommentCountByRepoID failed: %s", repo.Name) | |||
| log.Error("failed statistic: %s", repo.Name) | |||
| continue | |||
| log.Error("GetCommentCountByRepoID failed(%s): %v", repo.Name, err) | |||
| } | |||
| beginTime, endTime := getStatTime(date) | |||
| numVisits, err := repository.AppointProjectView(repo.OwnerName, repo.Name, beginTime, endTime) | |||
| var numVisits int | |||
| numVisits, err = repository.AppointProjectView(repo.OwnerName, repo.Name, beginTime, endTime) | |||
| if err != nil { | |||
| log.Error("Get numVisits failed", err) | |||
| numVisits = 0 | |||
| log.Error("AppointProjectView failed(%s): %v", repo.Name, err) | |||
| } | |||
| repoStat := models.RepoStatistic{ | |||
| RepoID: repo.ID, | |||
| Date: date, | |||
| NumWatches: int64(repo.NumWatches), | |||
| NumStars: int64(repo.NumStars), | |||
| NumDownloads: repo.CloneCnt, | |||
| NumComments: numComments, | |||
| NumVisits: int64(numVisits), | |||
| NumClosedIssues: int64(repo.NumClosedIssues), | |||
| NumVersions: numVersions, | |||
| NumDevMonths: repoGitStat.DevelopAge, | |||
| RepoSize: repo.Size, | |||
| DatasetSize: datasetSize, | |||
| NumModels: 0, | |||
| NumWikiViews: repoGitStat.WikiPages, | |||
| NumCommits: repo.NumCommit, | |||
| NumIssues: int64(repo.NumIssues), | |||
| NumPulls: int64(repo.NumPulls), | |||
| IssueFixedRate: issueFixedRate, | |||
| NumContributor: repoGitStat.Contributors, | |||
| NumKeyContributor: repoGitStat.KeyContributors, | |||
| RepoID: repo.ID, | |||
| Date: date, | |||
| NumWatches: int64(repo.NumWatches), | |||
| NumStars: int64(repo.NumStars), | |||
| NumDownloads: repo.CloneCnt, | |||
| NumComments: numComments, | |||
| NumVisits: int64(numVisits), | |||
| NumClosedIssues: int64(repo.NumClosedIssues), | |||
| NumVersions: numVersions, | |||
| NumDevMonths: numDevMonths, | |||
| RepoSize: repo.Size, | |||
| DatasetSize: datasetSize, | |||
| NumModels: 0, | |||
| NumWikiViews: numWikiViews, | |||
| NumCommits: repo.NumCommit, | |||
| NumIssues: int64(repo.NumIssues), | |||
| NumPulls: int64(repo.NumPulls), | |||
| IssueFixedRate: issueFixedRate, | |||
| NumContributor: numContributor, | |||
| NumKeyContributor: numKeyContributor, | |||
| NumCommitsGrowth: numCommitsGrowth, | |||
| NumCommitLinesGrowth: numCommitLinesGrowth, | |||
| NumContributorsGrowth: numContributorsGrowth, | |||
| } | |||
| dayBeforeDate := t.AddDate(0, 0, -1).Format("2006-01-02") | |||
| repoStatisticsBefore, err := models.GetRepoStatisticByDate(dayBeforeDate) | |||
| if err != nil { | |||
| log.Error("get data of day before the date failed ", err) | |||
| } else { | |||
| if len(repoStatisticsBefore) > 0 { | |||
| repoStatisticBefore := repoStatisticsBefore[0] | |||
| repoStat.NumWatchesAdded = repoStat.NumWatches - repoStatisticBefore.NumWatches | |||
| repoStat.NumStarsAdded = repoStat.NumStars - repoStatisticBefore.NumStars | |||
| repoStat.NumForksAdded = repoStat.NumForks - repoStatisticBefore.NumForks | |||
| repoStat.NumDownloadsAdded = repoStat.NumDownloads - repoStatisticBefore.NumDownloads | |||
| repoStat.NumCommentsAdded = repoStat.NumComments - repoStatisticBefore.NumComments | |||
| repoStat.NumClosedIssuesAdded = repoStat.NumClosedIssues - repoStatisticBefore.NumClosedIssues | |||
| repoStat.NumCommitsAdded = repoStat.NumCommits - repoStatisticBefore.NumCommits | |||
| repoStat.NumIssuesAdded = repoStat.NumIssues - repoStatisticBefore.NumIssues | |||
| repoStat.NumPullsAdded = repoStat.NumPulls - repoStatisticBefore.NumPulls | |||
| repoStat.NumContributorAdded = repoStat.NumContributor - repoStatisticBefore.NumContributor | |||
| } | |||
| } | |||
| day4MonthsAgo := t.AddDate(0, -4, 0) | |||
| repoStatisticFourMonthsAgo, err := models.GetOneRepoStatisticBeforeTime(day4MonthsAgo) | |||
| if err != nil { | |||
| log.Error("Get data of 4 moth ago failed.", err) | |||
| } else { | |||
| repoStat.NumCommentsGrowth = repoStat.NumComments - repoStatisticFourMonthsAgo.NumComments | |||
| repoStat.NumIssuesGrowth = repoStat.NumIssues - repoStatisticFourMonthsAgo.NumIssues | |||
| } | |||
| if _, err = models.InsertRepoStat(&repoStat); err != nil { | |||
| log.Error("InsertRepoStat failed: %s", repo.Name) | |||
| log.Error("InsertRepoStat failed(%s): %v", repo.Name, err) | |||
| log.Error("failed statistic: %s", repo.Name) | |||
| continue | |||
| } | |||
| tempRepoStat := models.RepoStatistic{ | |||
| RepoID: repoStat.RepoID, | |||
| Date: repoStat.Date, | |||
| Impact: normalization.GetImpactInitValue(repoStat.NumWatches, repoStat.NumStars, repoStat.NumForks, repoStat.NumDownloads, repoStat.NumComments, repoStat.NumVisits), | |||
| Completeness: normalization.GetCompleteInitValue(repoStat.NumClosedIssues, repoStat.NumVersions, repoStat.NumDevMonths, repoStat.DatasetSize, repoStat.NumModels, repoStat.NumWikiViews), | |||
| Liveness: normalization.GetLivenessInitValue(repoStat.NumCommits, repoStat.NumIssues, repoStat.NumPulls, repoStat.NumVisits), | |||
| ProjectHealth: normalization.GetProjectHealthInitValue(repoStat.IssueFixedRate), | |||
| TeamHealth: normalization.GetTeamHealthInitValue(repoStat.NumContributor, repoStat.NumKeyContributor, repoStat.NumContributorsGrowth), | |||
| Growth: normalization.GetRepoGrowthInitValue(repoStat.NumCommitLinesGrowth, repoStat.NumIssuesGrowth, repoStat.NumCommitsGrowth, repoStat.NumContributorsGrowth, repoStat.NumCommentsGrowth), | |||
| } | |||
| reposRadar = append(reposRadar, &tempRepoStat) | |||
| if i == 0 { | |||
| minRepoRadar = tempRepoStat | |||
| maxRepoRadar = tempRepoStat | |||
| } else { | |||
| if tempRepoStat.Impact < minRepoRadar.Impact { | |||
| minRepoRadar.Impact = tempRepoStat.Impact | |||
| } | |||
| if tempRepoStat.Impact > maxRepoRadar.Impact { | |||
| maxRepoRadar.Impact = tempRepoStat.Impact | |||
| } | |||
| if tempRepoStat.Completeness < minRepoRadar.Completeness { | |||
| minRepoRadar.Completeness = tempRepoStat.Completeness | |||
| } | |||
| if tempRepoStat.Completeness > maxRepoRadar.Completeness { | |||
| maxRepoRadar.Completeness = tempRepoStat.Completeness | |||
| } | |||
| if tempRepoStat.Liveness < minRepoRadar.Completeness { | |||
| minRepoRadar.Liveness = tempRepoStat.Liveness | |||
| } | |||
| if tempRepoStat.Liveness > maxRepoRadar.Liveness { | |||
| maxRepoRadar.Liveness = tempRepoStat.Liveness | |||
| } | |||
| if tempRepoStat.ProjectHealth < minRepoRadar.ProjectHealth { | |||
| minRepoRadar.ProjectHealth = tempRepoStat.ProjectHealth | |||
| } | |||
| if tempRepoStat.ProjectHealth > maxRepoRadar.ProjectHealth { | |||
| maxRepoRadar.ProjectHealth = tempRepoStat.ProjectHealth | |||
| } | |||
| if tempRepoStat.TeamHealth < minRepoRadar.TeamHealth { | |||
| minRepoRadar.TeamHealth = tempRepoStat.TeamHealth | |||
| } | |||
| if tempRepoStat.TeamHealth > maxRepoRadar.TeamHealth { | |||
| maxRepoRadar.TeamHealth = tempRepoStat.TeamHealth | |||
| } | |||
| if tempRepoStat.Growth < minRepoRadar.Growth { | |||
| minRepoRadar.Growth = tempRepoStat.Growth | |||
| } | |||
| if tempRepoStat.Growth > maxRepoRadar.Growth { | |||
| maxRepoRadar.Growth = tempRepoStat.Growth | |||
| } | |||
| } | |||
| log.Info("finish statistic: %s", repo.Name) | |||
| } | |||
| //radar map | |||
| log.Info("begin statistic radar") | |||
| for _, radarInit := range reposRadar { | |||
| radarInit.Impact = normalization.Normalization(radarInit.Impact, minRepoRadar.Impact, maxRepoRadar.Impact) | |||
| radarInit.Completeness = normalization.Normalization(radarInit.Completeness, minRepoRadar.Completeness, maxRepoRadar.Completeness) | |||
| radarInit.Liveness = normalization.Normalization(radarInit.Liveness, minRepoRadar.Liveness, maxRepoRadar.Liveness) | |||
| radarInit.ProjectHealth = normalization.Normalization(radarInit.ProjectHealth, minRepoRadar.ProjectHealth, maxRepoRadar.ProjectHealth) | |||
| radarInit.TeamHealth = normalization.Normalization(radarInit.TeamHealth, minRepoRadar.TeamHealth, maxRepoRadar.TeamHealth) | |||
| radarInit.Growth = normalization.Normalization(radarInit.Growth, minRepoRadar.Growth, maxRepoRadar.Growth) | |||
| radarInit.RadarTotal = normalization.GetRadarValue(radarInit.Impact, radarInit.Completeness, radarInit.Liveness, radarInit.ProjectHealth, radarInit.TeamHealth, radarInit.Growth) | |||
| models.UpdateRepoStat(radarInit) | |||
| } | |||
| log.Info("finish statistic: radar") | |||
| } | |||
| func getDatasetSize(repo *models.Repository) (int64, error) { | |||
| @@ -0,0 +1,94 @@ | |||
| package repo | |||
| import ( | |||
| "time" | |||
| "code.gitea.io/gitea/models" | |||
| "code.gitea.io/gitea/modules/log" | |||
| ) | |||
| func SummaryStatistic() { | |||
| log.Info("Generate summary statistic begin") | |||
| yesterday := time.Now().AddDate(0, 0, -1).Format("2006-01-02") | |||
| SummaryStatisticDaily(yesterday) | |||
| log.Info("Generate summary statistic end") | |||
| } | |||
| func SummaryStatisticDaily(date string) { | |||
| log.Info("%s", date) | |||
| if err := models.DeleteSummaryStatisticDaily(date); err != nil { | |||
| log.Error("DeleteRepoStatDaily failed: %v", err.Error()) | |||
| return | |||
| } | |||
| //user number | |||
| userNumber, err := models.GetUsersCount() | |||
| if err != nil { | |||
| log.Error("can not get user number", err) | |||
| userNumber = 0 | |||
| } | |||
| //organization number | |||
| organizationNumber, err := models.GetOrganizationsCount() | |||
| if err != nil { | |||
| log.Error("can not get orgnazition number", err) | |||
| organizationNumber = 0 | |||
| } | |||
| // repository number | |||
| repositoryNumer, err := models.GetAllRepositoriesCount() | |||
| if err != nil { | |||
| log.Error("can not get repository number", err) | |||
| repositoryNumer = 0 | |||
| } | |||
| //repository size | |||
| repositorySize, err := models.GetAllRepositoriesSize() | |||
| if err != nil { | |||
| log.Error("can not get repository size", err) | |||
| repositorySize = 0 | |||
| } | |||
| // dataset size | |||
| allDatasetSize, err := models.GetAllAttachmentSize() | |||
| if err != nil { | |||
| log.Error("can not get dataset size", err) | |||
| allDatasetSize = 0 | |||
| } | |||
| //topic repo number | |||
| topics, err := models.GetAllUsedTopics() | |||
| if err != nil { | |||
| log.Error("can not get topics", err) | |||
| } | |||
| var topicsCount [11]int | |||
| for _, topic := range topics { | |||
| index, exists := models.DomainMap[topic.Name] | |||
| if exists { | |||
| topicsCount[index] = topic.RepoCount | |||
| } | |||
| } | |||
| summaryStat := models.SummaryStatistic{ | |||
| Date: date, | |||
| NumUsers: userNumber, | |||
| RepoSize: repositorySize, | |||
| DatasetSize: allDatasetSize, | |||
| NumOrganizations: organizationNumber, | |||
| NumRepos: repositoryNumer, | |||
| NumRepoBigModel: topicsCount[0], | |||
| NumRepoAI: topicsCount[1], | |||
| NumRepoVision: topicsCount[2], | |||
| NumRepoNLP: topicsCount[3], | |||
| NumRepoML: topicsCount[4], | |||
| NumRepoNN: topicsCount[5], | |||
| NumRepoAutoDrive: topicsCount[6], | |||
| NumRepoRobot: topicsCount[7], | |||
| NumRepoLeagueLearn: topicsCount[8], | |||
| NumRepoDataMining: topicsCount[9], | |||
| NumRepoRISC: topicsCount[10], | |||
| } | |||
| if _, err = models.InsertSummaryStatistic(&summaryStat); err != nil { | |||
| log.Error("Insert summary Stat failed: %v", err.Error()) | |||
| } | |||
| log.Info("finish summary statistic") | |||
| } | |||
| @@ -6,13 +6,15 @@ package routes | |||
| import ( | |||
| "bytes" | |||
| "code.gitea.io/gitea/routers/operation" | |||
| "encoding/gob" | |||
| "net/http" | |||
| "path" | |||
| "text/template" | |||
| "time" | |||
| "code.gitea.io/gitea/routers/operation" | |||
| "code.gitea.io/gitea/routers/private" | |||
| "code.gitea.io/gitea/routers/secure" | |||
| "code.gitea.io/gitea/models" | |||
| @@ -33,7 +35,6 @@ import ( | |||
| "code.gitea.io/gitea/routers/dev" | |||
| "code.gitea.io/gitea/routers/events" | |||
| "code.gitea.io/gitea/routers/org" | |||
| "code.gitea.io/gitea/routers/private" | |||
| "code.gitea.io/gitea/routers/repo" | |||
| "code.gitea.io/gitea/routers/user" | |||
| userSetting "code.gitea.io/gitea/routers/user/setting" | |||
| @@ -113,14 +114,14 @@ func RouterHandler(level log.Level) func(ctx *macaron.Context) { | |||
| } | |||
| // SetLogMsgID set msgID in Context | |||
| func SetLogMsgID() func(ctx *macaron.Context) { | |||
| func SetLogMsgID() macaron.Handler { | |||
| return func(ctx *macaron.Context) { | |||
| start := time.Now() | |||
| uuid := gouuid.NewV4().String() | |||
| ctx.Data["MsgID"] = uuid | |||
| log.Info("Started %s %s for %s", log.ColoredMethod(ctx.Req.Method), ctx.Req.URL.RequestURI(), ctx.RemoteAddr(), ctx.Data["MsgID"]) | |||
| log.Info("%s Started %s %s for %s", ctx.Data["SignedUserName"], log.ColoredMethod(ctx.Req.Method), ctx.Req.URL.RequestURI(), ctx.RemoteAddr(), ctx.Data["MsgID"]) | |||
| rw := ctx.Resp.(macaron.ResponseWriter) | |||
| ctx.Next() | |||
| @@ -148,7 +149,7 @@ func NewMacaron() *macaron.Macaron { | |||
| m.Use(macaron.Logger()) | |||
| } | |||
| } | |||
| m.Use(SetLogMsgID()) | |||
| //m.Use(SetLogMsgID()) | |||
| // Access Logger is similar to Router Log but more configurable and by default is more like the NCSA Common Log format | |||
| if setting.EnableAccessLog { | |||
| setupAccessLogger(m) | |||
| @@ -256,6 +257,7 @@ func NewMacaron() *macaron.Macaron { | |||
| DisableDebug: !setting.EnablePprof, | |||
| })) | |||
| m.Use(context.Contexter()) | |||
| m.Use(SetLogMsgID()) | |||
| // OK we are now set-up enough to allow us to create a nicer recovery than | |||
| // the default macaron recovery | |||
| m.Use(context.Recovery()) | |||
| @@ -565,6 +567,8 @@ func RegisterRoutes(m *macaron.Macaron) { | |||
| m.Post("/private", repo.UpdatePublicAttachment) | |||
| m.Get("/get_chunks", repo.GetSuccessChunks) | |||
| m.Get("/new_multipart", repo.NewMultipart) | |||
| m.Put("/obs_proxy_multipart", repo.PutOBSProxyUpload) | |||
| m.Get("/obs_proxy_download", repo.GetOBSProxyDownload) | |||
| m.Get("/get_multipart_url", repo.GetMultipartUploadUrl) | |||
| m.Post("/complete_multipart", repo.CompleteMultipart) | |||
| m.Post("/update_chunk", repo.UpdateMultipart) | |||
| @@ -29,8 +29,12 @@ | |||
| {{.Repo.OwnerName}} / {{.Title}} | |||
| </a> | |||
| <div class="ui right metas"> | |||
| <span class="text grey">{{svg "octicon-tasklist" 16}} {{$.i18n.Tr (printf "dataset.task.%s" .Task)}}</span> | |||
| <span class="text grey">{{svg "octicon-tag" 16}}{{$.i18n.Tr (printf "dataset.category.%s" .Category)}}</span> | |||
| {{if .Task}} | |||
| <span class="text grey">{{svg "octicon-tasklist" 16}} {{$.i18n.Tr (printf "dataset.task.%s" .Task)}}</span> | |||
| {{end}} | |||
| {{if .Category}} | |||
| <span class="text grey">{{svg "octicon-tag" 16}}{{$.i18n.Tr (printf "dataset.category.%s" .Category)}}</span> | |||
| {{end}} | |||
| <span class="text grey">{{svg "octicon-flame" 16}} {{.DownloadTimes}}</span> | |||
| </div> | |||
| </div> | |||
| @@ -337,9 +337,9 @@ | |||
| 调试 | |||
| </a> | |||
| <form id="stopForm-{{.JobID}}" action="{{if eq .Status "STOPPED"}}javascript:void(0){{else}}{{$.Link}}/{{.JobID}}/stop{{end}}" method="post" style="margin-left:-1px;"> | |||
| <form id="stopForm-{{.JobID}}" action="{{if or (eq .Status "STOPPED") (eq .Status "FAILED")}}javascript:void(0){{else}}{{$.Link}}/{{.JobID}}/stop{{end}}" method="post" style="margin-left:-1px;"> | |||
| {{$.CsrfTokenHtml}} | |||
| <a class="ui basic {{if eq .Status "STOPPED"}}disabled {{else}}blue {{end}}button" onclick="document.getElementById('stopForm-{{.JobID}}').submit();"> | |||
| <a class="ui basic {{if or (eq .Status "STOPPED") (eq .Status "FAILED")}}disabled {{else}}blue {{end}}button" onclick="document.getElementById('stopForm-{{.JobID}}').submit();"> | |||
| 停止 | |||
| </a> | |||
| </form> | |||
| @@ -120,7 +120,7 @@ | |||
| <div class="repository new repo ui middle very relaxed page grid"> | |||
| <div class="column"> | |||
| {{template "base/alert" .}} | |||
| <div class="ui positive message" id="messageInfo"> | |||
| <div class="ui negative message" id="messageInfo"> | |||
| <p></p> | |||
| </div> | |||
| <form id="form_id" class="ui form" action="{{.Link}}" method="post"> | |||
| @@ -259,11 +259,11 @@ | |||
| let value_task = $("input[name='job_name']").val() | |||
| let value_image = $("input[name='image']").val() | |||
| let value_data = $("input[name='attachment']").val() | |||
| let re = /^[a-z0-9][a-z0-9-_]{1,35}[^_]$/ | |||
| let re = /^[a-z0-9][a-z0-9-_]{1,34}[a-z0-9-]$/ | |||
| let flag = re.test(value_task) | |||
| if(!flag){ | |||
| $('#messageInfo').css('display','block') | |||
| let str = '只能以小写字母或数字开头且只包含小写字母、数字、_和-、最长36个字符,不能下划线(-)结尾。' | |||
| let str = '只能以小写字母或数字开头且只包含小写字母、数字、_和-,不能以_结尾,最长36个字符。' | |||
| $('#messageInfo p').text(str) | |||
| return false | |||
| } | |||
| @@ -17,7 +17,7 @@ | |||
| <span class="ui basic basic button clipboard" data-clipboard-text="{{.DownloadURL}}" data-tooltip='{{$.i18n.Tr "dataset.copy_url"}}' data-clipboard-action="copy"{{if ne $.Type 0}} style="display:none;"{{end}}>{{svg "octicon-file" 16}}</span> | |||
| <span class="ui basic basic button clipboard" data-clipboard-text="{{.FileChunk.Md5}}" data-tooltip='{{$.i18n.Tr "dataset.copy_md5"}}' data-clipboard-action="copy">{{svg "octicon-file-binary" 16}}</span> | |||
| </div> | |||
| {{if ne .DecompressState 0}} | |||
| {{if eq .DecompressState 1}} | |||
| <div class="ui left mini icon buttons"> | |||
| <a class="ui basic blue button" href="datasets/dirs/{{.UUID}}?type={{$.Type}}" data-tooltip='{{$.i18n.Tr "dataset.directory"}}'>{{svg "octicon-file-directory" 16}}</a> | |||
| {{if $.IsSigned}} | |||
| @@ -123,7 +123,7 @@ | |||
| <select name="pre_predict_task" id="dataset_list_auto" onchange="dataset_auto_sele_Change(this)"> | |||
| {{if .Attachments}} | |||
| {{range .Attachments}} | |||
| <option value="{{.UUID}}">{{.Name}}</option> | |||
| <option value="{{.UUID}}">{{.Name}}</option> | |||
| {{end}} | |||
| {{end}} | |||
| </select> | |||
| @@ -100,7 +100,7 @@ | |||
| <div class="repository new repo ui middle very relaxed page grid"> | |||
| <div class="column"> | |||
| {{template "base/alert" .}} | |||
| <div class="ui positive message" id="messageInfo"> | |||
| <div class="ui negative message" id="messageInfo"> | |||
| <p></p> | |||
| </div> | |||
| <form class="ui form" id="form_id" action="{{.Link}}" method="post"> | |||
| @@ -11,16 +11,16 @@ | |||
| <div class="icon-wrapper"> | |||
| <i style="line-height: 1.5;color: #303643;font-weight: 900;" v-if="showInitTopic[i]" class="el-icon-check" ></i> | |||
| </div> | |||
| <div class="text">{{arr.topic_name}} </div> | |||
| <div class="text">{{arr.topic_name.toLowerCase()}} </div> | |||
| </div> | |||
| <div v-if="showInputValue" class="addition item-text" @click="postTopic"> | |||
| 点击或回车添加<b class="user-add-label-text">{{input}}</b>标签 | |||
| 点击或回车添加<b class="user-add-label-text">{{input.toLowerCase()}}</b>标签 | |||
| </div> | |||
| <div v-if="showAddTopic" class="item-text" @click="addPostTopic"> | |||
| <div class="icon-wrapper"> | |||
| <i style="line-height: 1.5;color: #303643;font-weight: 900;" v-if="showAddFlage" class="el-icon-check" ></i> | |||
| </div> | |||
| <div class="text">{{input}}</div> | |||
| <div class="text">{{input.toLowerCase()}}</div> | |||
| </div> | |||
| </div> | |||
| @@ -134,7 +134,7 @@ export default { | |||
| this.showSearchTopic = true | |||
| } | |||
| else if(this.arrayTopics.indexOf(this.input)>-1){ | |||
| else if(this.arrayTopics.indexOf(this.input.toLowerCase())>-1){ | |||
| this.showInputValue = false | |||
| this.showSearchTopic = false | |||
| @@ -142,7 +142,7 @@ export default { | |||
| this.showInitTopic = [] | |||
| let timestamp=new Date().getTime() | |||
| this.params.q = this.input | |||
| this.params.q = this.input.toLowerCase() | |||
| this.params._ = timestamp | |||
| this.$axios.get('/api/v1/topics/search',{ | |||
| params:this.params | |||
| @@ -164,7 +164,7 @@ export default { | |||
| let findelement = this.array.some((item)=>{ | |||
| return item.topic_name===this.input | |||
| return item.topic_name===this.input.toLowerCase() | |||
| }) | |||
| this.showInputValue = !findelement | |||
| @@ -224,11 +224,11 @@ export default { | |||
| return | |||
| }else{ | |||
| let topic = this.input | |||
| if(this.arrayTopics.includes(topic)){ | |||
| if(this.arrayTopics.includes(topic.toLowerCase())){ | |||
| return | |||
| } | |||
| else{ | |||
| this.arrayTopics.push(topic) | |||
| this.arrayTopics.push(topic.toLowerCase()) | |||
| let topics = this.arrayTopics | |||
| let strTopics = topics.join(',') | |||
| @@ -250,7 +250,10 @@ export default { | |||
| addPostTopic(){ | |||
| if(this.showAddFlage){ | |||
| this.arrayTopics.pop() | |||
| // this.arrayTopics.pop() | |||
| let cancleIndex = this.arrayTopics.indexOf(this.input) | |||
| this.arrayTopics.splice(cancleIndex,1) | |||
| let topics = this.arrayTopics | |||
| let strTopics = topics.join(',') | |||
| let data = this.qs.stringify({ | |||
| @@ -268,7 +271,7 @@ export default { | |||
| } | |||
| else if(!this.showAddFlage){ | |||
| let topic = this.input | |||
| this.arrayTopics.push(topic) | |||
| this.arrayTopics.push(topic.toLowerCase()) | |||
| let topics = this.arrayTopics | |||
| let strTopics = topics.join(',') | |||