chore(controller): Fix modernize linter (#26303)

Signed-off-by: Matthieu MOREL <matthieu.morel35@gmail.com>
This commit is contained in:
Matthieu MOREL
2026-02-06 23:16:37 +01:00
committed by GitHub
parent 6e0c949dd2
commit e2f7e7d27f
7 changed files with 21 additions and 26 deletions

View File

@@ -5,6 +5,7 @@ import (
"encoding/json"
stderrors "errors"
"fmt"
"maps"
"math"
"math/rand"
"net/http"
@@ -927,14 +928,14 @@ func (ctrl *ApplicationController) Run(ctx context.Context, statusProcessors int
go func() { errors.CheckError(ctrl.stateCache.Run(ctx)) }()
go func() { errors.CheckError(ctrl.metricsServer.ListenAndServe()) }()
for i := 0; i < statusProcessors; i++ {
for range statusProcessors {
go wait.Until(func() {
for ctrl.processAppRefreshQueueItem() {
}
}, time.Second, ctx.Done())
}
for i := 0; i < operationProcessors; i++ {
for range operationProcessors {
go wait.Until(func() {
for ctrl.processAppOperationQueueItem() {
}
@@ -2131,9 +2132,7 @@ func (ctrl *ApplicationController) persistAppStatus(orig *appv1.Application, new
var newAnnotations map[string]string
if orig.GetAnnotations() != nil {
newAnnotations = make(map[string]string)
for k, v := range orig.GetAnnotations() {
newAnnotations[k] = v
}
maps.Copy(newAnnotations, orig.GetAnnotations())
delete(newAnnotations, appv1.AnnotationKeyRefresh)
delete(newAnnotations, appv1.AnnotationKeyHydrate)
}
@@ -2374,7 +2373,7 @@ func (ctrl *ApplicationController) selfHealRemainingBackoff(app *appv1.Applicati
backOff.Steps = selfHealAttemptsCount
var delay time.Duration
steps := backOff.Steps
for i := 0; i < steps; i++ {
for range steps {
delay = backOff.Step()
}
if timeSinceOperation == nil {

View File

@@ -4,6 +4,7 @@ import (
"context"
"errors"
"fmt"
"maps"
"math"
"net"
"net/url"
@@ -366,9 +367,7 @@ func getAppRecursive(r *clustercache.Resource, ns map[kube.ResourceKey]*clusterc
gv := ownerRefGV(ownerRef)
if parent, ok := ns[kube.NewResourceKey(gv.Group, ownerRef.Kind, r.Ref.Namespace, ownerRef.Name)]; ok {
visitedBranch := make(map[kube.ResourceKey]bool, len(visited))
for k, v := range visited {
visitedBranch[k] = v
}
maps.Copy(visitedBranch, visited)
app, ok := getAppRecursive(parent, ns, visitedBranch)
if app != "" || !ok {
return app, ok

View File

@@ -410,7 +410,6 @@ argocd_app_labels{label_non_existing="",name="my-app-3",namespace="argocd",proje
}
for _, c := range cases {
c := c
t.Run(c.description, func(t *testing.T) {
testMetricServer(t, c.applications, c.responseContains, c.metricLabels, []string{})
})
@@ -464,7 +463,6 @@ argocd_app_condition{condition="ExcludedResourceWarning",name="my-app-4",namespa
}
for _, c := range cases {
c := c
t.Run(c.description, func(t *testing.T) {
testMetricServer(t, c.applications, c.responseContains, []string{}, c.metricConditions)
})
@@ -506,7 +504,7 @@ argocd_app_sync_total{dest_server="https://localhost:6443",dry_run="false",name=
// assertMetricsPrinted asserts every line in the expected lines appears in the body
func assertMetricsPrinted(t *testing.T, expectedLines, body string) {
t.Helper()
for _, line := range strings.Split(expectedLines, "\n") {
for line := range strings.SplitSeq(expectedLines, "\n") {
if line == "" {
continue
}
@@ -517,7 +515,7 @@ func assertMetricsPrinted(t *testing.T, expectedLines, body string) {
// assertMetricsNotPrinted
func assertMetricsNotPrinted(t *testing.T, expectedLines, body string) {
t.Helper()
for _, line := range strings.Split(expectedLines, "\n") {
for line := range strings.SplitSeq(expectedLines, "\n") {
if line == "" {
continue
}

View File

@@ -1,6 +1,7 @@
package sharding
import (
"maps"
"sync"
log "github.com/sirupsen/logrus"
@@ -134,9 +135,7 @@ func (sharding *ClusterSharding) GetDistribution() map[string]int {
shards := sharding.Shards
distribution := make(map[string]int, len(shards))
for k, v := range shards {
distribution[k] = v
}
maps.Copy(distribution, shards)
return distribution
}

View File

@@ -224,7 +224,7 @@ func TestGetShardByIndexModuloReplicasCountDistributionFunctionWhenClusterNumber
// and for 4096 clusters, execution time was under 9s
// The other implementation was giving almost linear time of 400ms up to 10'000 clusters
clusterPointers := []*v1alpha1.Cluster{}
for i := 0; i < 2048; i++ {
for i := range 2048 {
cluster := createCluster(fmt.Sprintf("cluster-%d", i), strconv.Itoa(i))
clusterPointers = append(clusterPointers, &cluster)
}
@@ -282,7 +282,7 @@ func TestConsistentHashingWhenClusterIsAddedAndRemoved(t *testing.T) {
prefix := "cluster"
clusters := []v1alpha1.Cluster{}
for i := 0; i < clusterCount; i++ {
for i := range clusterCount {
id := fmt.Sprintf("%06d", i)
cluster := fmt.Sprintf("%s-%s", prefix, id)
clusters = append(clusters, createCluster(cluster, id))
@@ -298,7 +298,7 @@ func TestConsistentHashingWhenClusterIsAddedAndRemoved(t *testing.T) {
assert.Equal(t, 0, distributionFunction(nil))
distributionMap := map[int]int{}
assignementMap := map[string]int{}
for i := 0; i < clusterCount; i++ {
for i := range clusterCount {
assignedShard := distributionFunction(&clusters[i])
assignementMap[clusters[i].ID] = assignedShard
distributionMap[assignedShard]++
@@ -330,7 +330,7 @@ func TestConsistentHashingWhenClusterIsAddedAndRemoved(t *testing.T) {
replicasCount = 2
distributionFunction = ConsistentHashingWithBoundedLoadsDistributionFunction(getClusterAccessor(clusterList.Items), appAccessor, replicasCount)
removedCluster := clusterList.Items[len(clusterList.Items)-1]
for i := 0; i < clusterCount; i++ {
for i := range clusterCount {
c := &clusters[i]
assignedShard := distributionFunction(c)
prevıouslyAssignedShard := assignementMap[clusters[i].ID]

View File

@@ -1,6 +1,8 @@
package controller
import (
"maps"
gitopscommon "github.com/argoproj/gitops-engine/pkg/sync/common"
"k8s.io/apimachinery/pkg/apis/meta/v1/unstructured"
@@ -47,9 +49,7 @@ func syncNamespace(syncPolicy *v1alpha1.SyncPolicy) func(m *unstructured.Unstruc
// with server-side apply
func appendSSAAnnotation(in map[string]string) map[string]string {
r := map[string]string{}
for k, v := range in {
r[k] = v
}
maps.Copy(r, in)
r[gitopscommon.AnnotationSyncOptions] = gitopscommon.SyncOptionServerSideApply
return r
}

View File

@@ -18,9 +18,9 @@ func TestGenerate(t *testing.T) {
globalCount.Store(0)
// Run goroutines in parallel to test for race conditions
for g := 0; g < goroutines; g++ {
for range goroutines {
go func() {
for i := 0; i < idsPerGoroutine; i++ {
for range idsPerGoroutine {
id, err := Generate()
if err != nil {
errCh <- err
@@ -32,7 +32,7 @@ func TestGenerate(t *testing.T) {
}
ids := make(map[string]any)
for i := 0; i < goroutines*idsPerGoroutine; i++ {
for range goroutines * idsPerGoroutine {
select {
case err := <-errCh:
require.NoError(t, err)