You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

chunked.go 3.8 kB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155
  1. package ops2
  2. import (
  3. "context"
  4. "fmt"
  5. "io"
  6. "github.com/samber/lo"
  7. "gitlink.org.cn/cloudream/common/pkgs/future"
  8. "gitlink.org.cn/cloudream/common/pkgs/ioswitch/dag"
  9. "gitlink.org.cn/cloudream/common/pkgs/ioswitch/exec"
  10. "gitlink.org.cn/cloudream/common/pkgs/ioswitch/utils"
  11. "gitlink.org.cn/cloudream/common/utils/io2"
  12. "gitlink.org.cn/cloudream/storage/common/pkgs/ioswitch2"
  13. "golang.org/x/sync/semaphore"
  14. )
  15. func init() {
  16. exec.UseOp[*ChunkedSplit]()
  17. exec.UseOp[*ChunkedJoin]()
  18. }
  19. type ChunkedSplit struct {
  20. Input *exec.StreamVar `json:"input"`
  21. Outputs []*exec.StreamVar `json:"outputs"`
  22. ChunkSize int `json:"chunkSize"`
  23. PaddingZeros bool `json:"paddingZeros"`
  24. }
  25. func (o *ChunkedSplit) Execute(ctx context.Context, e *exec.Executor) error {
  26. err := e.BindVars(ctx, o.Input)
  27. if err != nil {
  28. return err
  29. }
  30. defer o.Input.Stream.Close()
  31. outputs := io2.ChunkedSplit(o.Input.Stream, o.ChunkSize, len(o.Outputs), io2.ChunkedSplitOption{
  32. PaddingZeros: o.PaddingZeros,
  33. })
  34. sem := semaphore.NewWeighted(int64(len(outputs)))
  35. for i := range outputs {
  36. sem.Acquire(ctx, 1)
  37. o.Outputs[i].Stream = io2.AfterReadClosedOnce(outputs[i], func(closer io.ReadCloser) {
  38. sem.Release(1)
  39. })
  40. }
  41. exec.PutArrayVars(e, o.Outputs)
  42. return sem.Acquire(ctx, int64(len(outputs)))
  43. }
  44. func (o *ChunkedSplit) String() string {
  45. return fmt.Sprintf(
  46. "ChunkedSplit(chunkSize=%v, paddingZeros=%v), %v -> (%v)",
  47. o.ChunkSize,
  48. o.PaddingZeros,
  49. o.Input.ID,
  50. utils.FormatVarIDs(o.Outputs),
  51. )
  52. }
  53. type ChunkedJoin struct {
  54. Inputs []*exec.StreamVar `json:"inputs"`
  55. Output *exec.StreamVar `json:"output"`
  56. ChunkSize int `json:"chunkSize"`
  57. }
  58. func (o *ChunkedJoin) Execute(ctx context.Context, e *exec.Executor) error {
  59. err := exec.BindArrayVars(e, ctx, o.Inputs)
  60. if err != nil {
  61. return err
  62. }
  63. var strReaders []io.Reader
  64. for _, s := range o.Inputs {
  65. strReaders = append(strReaders, s.Stream)
  66. }
  67. defer func() {
  68. for _, str := range o.Inputs {
  69. str.Stream.Close()
  70. }
  71. }()
  72. fut := future.NewSetVoid()
  73. o.Output.Stream = io2.AfterReadClosedOnce(io2.BufferedChunkedJoin(strReaders, o.ChunkSize), func(closer io.ReadCloser) {
  74. fut.SetVoid()
  75. })
  76. e.PutVars(o.Output)
  77. return fut.Wait(ctx)
  78. }
  79. func (o *ChunkedJoin) String() string {
  80. return fmt.Sprintf(
  81. "ChunkedJoin(chunkSize=%v), (%v) -> %v",
  82. o.ChunkSize,
  83. utils.FormatVarIDs(o.Inputs),
  84. o.Output.ID,
  85. )
  86. }
  87. type ChunkedSplitType struct {
  88. OutputCount int
  89. ChunkSize int
  90. }
  91. func (t *ChunkedSplitType) InitNode(node *dag.Node) {
  92. dag.NodeDeclareInputStream(node, 1)
  93. for i := 0; i < t.OutputCount; i++ {
  94. dag.NodeNewOutputStream(node, &ioswitch2.VarProps{
  95. StreamIndex: i,
  96. })
  97. }
  98. }
  99. func (t *ChunkedSplitType) GenerateOp(op *dag.Node) (exec.Op, error) {
  100. return &ChunkedSplit{
  101. Input: op.InputStreams[0].Var,
  102. Outputs: lo.Map(op.OutputStreams, func(v *dag.StreamVar, idx int) *exec.StreamVar {
  103. return v.Var
  104. }),
  105. ChunkSize: t.ChunkSize,
  106. PaddingZeros: true,
  107. }, nil
  108. }
  109. func (t *ChunkedSplitType) String(node *dag.Node) string {
  110. return fmt.Sprintf("ChunkedSplit[%v]%v%v", t.ChunkSize, formatStreamIO(node), formatValueIO(node))
  111. }
  112. type ChunkedJoinType struct {
  113. InputCount int
  114. ChunkSize int
  115. }
  116. func (t *ChunkedJoinType) InitNode(node *dag.Node) {
  117. dag.NodeDeclareInputStream(node, t.InputCount)
  118. dag.NodeNewOutputStream(node, &ioswitch2.VarProps{})
  119. }
  120. func (t *ChunkedJoinType) GenerateOp(op *dag.Node) (exec.Op, error) {
  121. return &ChunkedJoin{
  122. Inputs: lo.Map(op.InputStreams, func(v *dag.StreamVar, idx int) *exec.StreamVar {
  123. return v.Var
  124. }),
  125. Output: op.OutputStreams[0].Var,
  126. ChunkSize: t.ChunkSize,
  127. }, nil
  128. }
  129. func (t *ChunkedJoinType) String(node *dag.Node) string {
  130. return fmt.Sprintf("ChunkedJoin[%v]%v%v", t.ChunkSize, formatStreamIO(node), formatValueIO(node))
  131. }

本项目旨在将云际存储公共基础设施化,使个人及企业可低门槛使用高效的云际存储服务(安装开箱即用云际存储客户端即可,无需关注其他组件的部署),同时支持用户灵活便捷定制云际存储的功能细节。