我已经将文件摄取到了德鲁伊中,极大地表明了摄取是成功的。但是,当我检查摄入报告时,所有行都进行了错误处理,但数据源显示在“数据源”选项卡中。

我试图将行从20M减少到仅20行。这是我的配置文件:

"type" : "index",
  "spec" : {
    "ioConfig" : {
      "type" : "index",
      "firehose" : {
        "type" : "local",
        "baseDir" : "/home/data/Salutica",
      "filter" : "outDashboard2RawV3.csv"
      }
    },
    "dataSchema" : {
        "dataSource": "DaTRUE2_Dashboard_V3",
      "granularitySpec" : {
        "type" : "uniform",
        "segmentGranularity" : "WEEK",
        "queryGranularity" : "none",
        "intervals" : ["2017-05-08/2019-05-17"],
        "rollup" : false
      },
      "parser" : {
        "type" : "string",
        "parseSpec": {
          "format" : "csv",
           "timestampSpec" : {
            "column" : "Date_Time",
            "format" : "auto"
         },
        "columns" : [
          "Main_ID","Parameter_ID","Date_Time","Serial_Number","Status","Station_ID",
          "Station_Type","Parameter_Name","Failed_Date_Time","Failed_Measurement",
          "Database_Name","Date_Time_Year","Date_Time_Month",
          "Date_Time_Day","Date_Time_Hour","Date_Time_Weekday","Status_New"
        ],
        "dimensionsSpec" : {
          "dimensions" : [
              "Date_Time","Serial_Number","Status","Station_ID",
              "Station_Type","Parameter_Name","Failed_Date_Time",
              "Failed_Measurement","Database_Name","Status_New",
              {
                  "name" : "Main_ID",
                  "type" : "long"
              },
              {
                  "name" : "Parameter_ID",
                  "type" : "long"
              },
              {
                  "name" : "Date_Time_Year",
                  "type" : "long"
              },
              {
                  "name" : "Date_Time_Month",
                  "type" : "long"
              },
              {
                  "name" : "Date_Time_Day",
                  "type" : "long"
              },
              {
                  "name" : "Date_Time_Hour",
                  "type" : "long"
              },
              {
                  "name" : "Date_Time_Weekday",
                  "type" : "long"
              }
                  ]
           }
        }
      },
      "metricsSpec" : [
        {
          "name" : "count",
          "type" : "count"
        }
      ]
    },
    "tuningConfig" : {
      "type" : "index",
      "partitionsSpec" : {
        "type" : "hashed",
        "targetPartitionSize" : 5000000
      },
      "jobProperties" : {}
    }
  }
}


报告:


{“ ingestionStatsAndErrors”:{“ taskId”:“ index_DaTRUE2_Dashboard_V3_2019-09-10T01:16:47.113Z”,“ payload”:{“ ingestionState”:“ COMPLETED”,“ unparseableEvents”:{},“ rowStats”:{“ determinePartitions “:{” processed“:0,” processedWithError“:0,” thrownAway“:0,” unparseable“:0},” buildSegments“:{” processed“:0,” processedWithError“:20606701,” thrownAway“:0 ,“ unparseable”:1}},“ errorMsg”:null},“ type”:“ ingestionStatsAndErrors”}}


我期望这样:


{“已处理”:20606701,“ processedWithError”:0,“ throwAway”:0,“不可解析”:1}},“ errorMsg”:null},“ type”:“ ingestionStatsAndErrors”}}


代替这个:


{“已处理”:0,“ processedWithError”:20606701,“ throwAway”:0,“ unparseable”:1}},“ errorMsg”:null},“ type”:“ ingestionStatsAndErrors”}}

最佳答案

以下是我从csv输入的数据;

    "Main_ID","Parameter_ID","Date_Time","Serial_Number","Status","Station_ID","Station_Type","Parameter_Name","Failed_Date_Time","Failed_Measurement","Database_Name","Date_Time_Year","Date_Time_Month","Date_Time_Day","Date_Time_Hour","Date_Time_Weekday","Status_New"
1,3,"2018-10-05 15:00:55","1840SDF00038","Passed","ST1","BLTBoard","1.8V","","","DaTRUE2Left",2018,10,5,15,"Friday","Passed"
1,4,"2018-10-05 15:00:55","1840SDF00038","Passed","ST1","BLTBoard","1.35V","","","DaTRUE2Left",2018,10,5,15,"Friday","Passed"
1,5,"2018-10-05 15:00:55","1840SDF00038","Passed","ST1","BLTBoard","Isc_VChrg","","","DaTRUE2Left",2018,10,5,15,"Friday","Passed"
1,6,"2018-10-05 15:00:55","1840SDF00038","Passed","ST1","BLTBoard","Isc_VBAT","","","DaTRUE2Left",2018,10,5,15,"Friday","Passed"

关于java - 报表任务管理器中的processedWithError是什么意思?,我们在Stack Overflow上找到一个类似的问题:https://stackoverflow.com/questions/57868698/

10-09 04:03