package shell import ( "context" "flag" "fmt" "github.com/seaweedfs/seaweedfs/weed/cluster" "github.com/seaweedfs/seaweedfs/weed/pb" "github.com/seaweedfs/seaweedfs/weed/pb/filer_pb" "github.com/seaweedfs/seaweedfs/weed/pb/volume_server_pb" "io" "github.com/seaweedfs/seaweedfs/weed/pb/master_pb" ) func init() { Commands = append(Commands, &commandClusterPs{}) } type commandClusterPs struct { } func (c *commandClusterPs) Name() string { return "cluster.ps" } func (c *commandClusterPs) Help() string { return `check current cluster process status cluster.ps ` } func (c *commandClusterPs) Do(args []string, commandEnv *CommandEnv, writer io.Writer) (err error) { clusterPsCommand := flag.NewFlagSet(c.Name(), flag.ContinueOnError) if err = clusterPsCommand.Parse(args); err != nil { return nil } var filerNodes []*master_pb.ListClusterNodesResponse_ClusterNode var mqBrokerNodes []*master_pb.ListClusterNodesResponse_ClusterNode // get the list of filers err = commandEnv.MasterClient.WithClient(false, func(client master_pb.SeaweedClient) error { resp, err := client.ListClusterNodes(context.Background(), &master_pb.ListClusterNodesRequest{ ClientType: cluster.FilerType, FilerGroup: *commandEnv.option.FilerGroup, }) if err != nil { return err } filerNodes = resp.ClusterNodes return err }) if err != nil { return } // get the list of message queue brokers err = commandEnv.MasterClient.WithClient(false, func(client master_pb.SeaweedClient) error { resp, err := client.ListClusterNodes(context.Background(), &master_pb.ListClusterNodesRequest{ ClientType: cluster.BrokerType, FilerGroup: *commandEnv.option.FilerGroup, }) if err != nil { return err } mqBrokerNodes = resp.ClusterNodes return err }) if err != nil { return } if len(mqBrokerNodes) > 0 { fmt.Fprintf(writer, "* message queue brokers %d\n", len(mqBrokerNodes)) for _, node := range mqBrokerNodes { fmt.Fprintf(writer, " * %s (%v)\n", node.Address, node.Version) if node.DataCenter != "" { fmt.Fprintf(writer, " DataCenter: %v\n", node.DataCenter) } if node.Rack != "" { fmt.Fprintf(writer, " Rack: %v\n", node.Rack) } if node.IsLeader { fmt.Fprintf(writer, " IsLeader: %v\n", true) } } } fmt.Fprintf(writer, "* filers %d\n", len(filerNodes)) for _, node := range filerNodes { fmt.Fprintf(writer, " * %s (%v)\n", node.Address, node.Version) if node.DataCenter != "" { fmt.Fprintf(writer, " DataCenter: %v\n", node.DataCenter) } if node.Rack != "" { fmt.Fprintf(writer, " Rack: %v\n", node.Rack) } pb.WithFilerClient(false, pb.ServerAddress(node.Address), commandEnv.option.GrpcDialOption, func(client filer_pb.SeaweedFilerClient) error { resp, err := client.GetFilerConfiguration(context.Background(), &filer_pb.GetFilerConfigurationRequest{}) if err == nil { if resp.FilerGroup != "" { fmt.Fprintf(writer, " filer group: %s\n", resp.FilerGroup) } fmt.Fprintf(writer, " signature: %d\n", resp.Signature) } else { fmt.Fprintf(writer, " failed to connect: %v\n", err) } return err }) } // collect volume servers var volumeServers []pb.ServerAddress t, _, err := collectTopologyInfo(commandEnv, 0) if err != nil { return err } for _, dc := range t.DataCenterInfos { for _, r := range dc.RackInfos { for _, dn := range r.DataNodeInfos { volumeServers = append(volumeServers, pb.NewServerAddressFromDataNode(dn)) } } } fmt.Fprintf(writer, "* volume servers %d\n", len(volumeServers)) for _, dc := range t.DataCenterInfos { fmt.Fprintf(writer, " * data center: %s\n", dc.Id) for _, r := range dc.RackInfos { fmt.Fprintf(writer, " * rack: %s\n", r.Id) for _, dn := range r.DataNodeInfos { pb.WithVolumeServerClient(false, pb.NewServerAddressFromDataNode(dn), commandEnv.option.GrpcDialOption, func(client volume_server_pb.VolumeServerClient) error { resp, err := client.VolumeServerStatus(context.Background(), &volume_server_pb.VolumeServerStatusRequest{}) if err == nil { fmt.Fprintf(writer, " * %s (%v)\n", dn.Id, resp.Version) } return err }) } } } return nil }