Upsert在Hudi中的实现分析

简介: Upsert在Hudi中的实现分析

介绍

Hudi支持Upsert语义,即将数据插入更新至Hudi数据集中,在借助索引机制完成数据查询后(查找记录位于哪个文件),再将该记录的位置信息回推至记录本身,然后对于已经存在于文件的记录使用UPDATE,而未存在于文件中的记录使用INSERT。本篇继续分析记录如何进行插入更新的。

分析

还是从HoodieBloomIndex#tagLocation开始进行分析,其核心代码如下。

public JavaRDD<HoodieRecord<T>> tagLocation(JavaRDD<HoodieRecord<T>> recordRDD, JavaSparkContext jsc,
     HoodieTable<T> hoodieTable) {
  ...
   // Lookup indexes for all the partition/recordkey pair
   JavaPairRDD<HoodieKey, HoodieRecordLocation> keyFilenamePairRDD =
       lookupIndex(partitionRecordKeyPairRDD, jsc, hoodieTable);
...
   JavaRDD<HoodieRecord<T>> taggedRecordRDD = tagLocationBacktoRecords(keyFilenamePairRDD, recordRDD);
...
   return taggedRecordRDD;
}

经过lookupIndex方法后只是找出了哪些记录存在于哪些文件,此时在原始记录中还并未有位置信息,需要经过tagLocationBacktoRecords将位置信息回推到记录中,该方法核心代码如下

protected JavaRDD<HoodieRecord<T>> tagLocationBacktoRecords(
     JavaPairRDD<HoodieKey, HoodieRecordLocation> keyFilenamePairRDD, JavaRDD<HoodieRecord<T>> recordRDD) {
   JavaPairRDD<HoodieKey, HoodieRecord<T>> keyRecordPairRDD =
       recordRDD.mapToPair(record -> new Tuple2<>(record.getKey(), record));
   // Here as the recordRDD might have more data than rowKeyRDD (some rowKeys' fileId is null),
   // so we do left outer join.
   return keyRecordPairRDD.leftOuterJoin(keyFilenamePairRDD).values()
      .map(v1 -> getTaggedRecord(v1._1, Option.ofNullable(v1._2.orNull())));
}

可以看到该方法的核心逻辑非常简单,先把最原始的记录进行一次变换(方便后续进行join操作),然后将变换的记录与之前已经查找的记录进行一次左外连接就完成了记录位置的回推操作(不得不感叹RDD太强大了)。

在完成位置信息回推后,就可以通过upsertRecordsInternal进行插入更新了,该方法核心代码如下

private JavaRDD<WriteStatus> upsertRecordsInternal(JavaRDD<HoodieRecord<T>> preppedRecords, String commitTime,
     HoodieTable<T> hoodieTable, final boolean isUpsert) {
  ...
   WorkloadProfile profile = null;
   if (hoodieTable.isWorkloadProfileNeeded()) {
     profile = new WorkloadProfile(preppedRecords);
     saveWorkloadProfileMetadataToInflight(profile, hoodieTable, commitTime);
  }
   // partition using the insert partitioner
   final Partitioner partitioner = getPartitioner(hoodieTable, isUpsert, profile);
   JavaRDD<HoodieRecord<T>> partitionedRecords = partition(preppedRecords, partitioner);
   JavaRDD<WriteStatus> writeStatusRDD = partitionedRecords.mapPartitionsWithIndex((partition, recordItr) -> {
     if (isUpsert) {
       return hoodieTable.handleUpsertPartition(commitTime, partition, recordItr, partitioner);
    } else {
       return hoodieTable.handleInsertPartition(commitTime, partition, recordItr, partitioner);
    }
  }, true).flatMap(List::iterator);
   return updateIndexAndCommitIfNeeded(writeStatusRDD, hoodieTable, commitTime);
}

首先会对记录进行统计,如本次处理中每个分区插入、更新多少条记录,然后根据不同的表类型(Merge On Read/Copy On Write)来获取对应的Partitioner进行重新分区,这里以HoodieCopyOnWriteTable$UpsertPartitioner为例进行分析。构造该对象时会利用profile信息来进行必要的初始化。

UpsertPartitioner(WorkloadProfile profile) {
    ...
     assignUpdates(profile);
     assignInserts(profile);
...
}
其中assignUpdates相应方法核心代码如下
   private void assignUpdates(WorkloadProfile profile) {
     // each update location gets a partition
     WorkloadStat gStat = profile.getGlobalStat();
     for (Map.Entry<String, Pair<String, Long>> updateLocEntry : gStat.getUpdateLocationToCount().entrySet()) {
       addUpdateBucket(updateLocEntry.getKey());
    }
  }
   private int addUpdateBucket(String fileIdHint) {
     int bucket = totalBuckets;
     updateLocationToBucket.put(fileIdHint, bucket);
     BucketInfo bucketInfo = new BucketInfo();
     bucketInfo.bucketType = BucketType.UPDATE;
     bucketInfo.fileIdPrefix = fileIdHint;
     bucketInfoMap.put(totalBuckets, bucketInfo);
     totalBuckets++;
     return bucket;
  }

该方法借助统计信息遍历需要更新的记录,并生成UPDATE类型的桶信息,文件名 -> 桶序号的映射、桶序号与桶信息的映射都会被保存。

assignInserts方法核心代码如下

private void assignInserts(WorkloadProfile profile) {
     // for new inserts, compute buckets depending on how many records we have for each partition
     Set<String> partitionPaths = profile.getPartitionPaths();
    ...
     for (String partitionPath : partitionPaths) {
       WorkloadStat pStat = profile.getWorkloadStat(partitionPath);
       if (pStat.getNumInserts() > 0) {
         List<SmallFile> smallFiles = getSmallFiles(partitionPath);
         long totalUnassignedInserts = pStat.getNumInserts();
         List<Integer> bucketNumbers = new ArrayList<>();
         List<Long> recordsPerBucket = new ArrayList<>();
         // first try packing this into one of the smallFiles
         for (SmallFile smallFile : smallFiles) {
           long recordsToAppend = Math.min((config.getParquetMaxFileSize() - smallFile.sizeBytes) / averageRecordSize,
                   totalUnassignedInserts);
           if (recordsToAppend > 0 && totalUnassignedInserts > 0) {
             // create a new bucket or re-use an existing bucket
             int bucket;
             if (updateLocationToBucket.containsKey(smallFile.location.getFileId())) {
               bucket = updateLocationToBucket.get(smallFile.location.getFileId());
            } else {
               bucket = addUpdateBucket(smallFile.location.getFileId());
            }
             bucketNumbers.add(bucket);
             recordsPerBucket.add(recordsToAppend);
             totalUnassignedInserts -= recordsToAppend;
          }
        }
         // if we have anything more, create new insert buckets, like normal
         if (totalUnassignedInserts > 0) {
           long insertRecordsPerBucket = config.getCopyOnWriteInsertSplitSize();
           if (config.shouldAutoTuneInsertSplits()) {
             insertRecordsPerBucket = config.getParquetMaxFileSize() / averageRecordSize;
          }
           int insertBuckets = (int) Math.ceil((1.0 * totalUnassignedInserts) / insertRecordsPerBucket);
           for (int b = 0; b < insertBuckets; b++) {
             bucketNumbers.add(totalBuckets);
             recordsPerBucket.add(totalUnassignedInserts / insertBuckets);
             BucketInfo bucketInfo = new BucketInfo();
             bucketInfo.bucketType = BucketType.INSERT;
             bucketInfo.fileIdPrefix = FSUtils.createNewFileIdPfx();
             bucketInfoMap.put(totalBuckets, bucketInfo);
             totalBuckets++;
          }
        }
         // Go over all such buckets, and assign weights as per amount of incoming inserts.
         List<InsertBucket> insertBuckets = new ArrayList<>();
         for (int i = 0; i < bucketNumbers.size(); i++) {
           InsertBucket bkt = new InsertBucket();
           bkt.bucketNumber = bucketNumbers.get(i);
           bkt.weight = (1.0 * recordsPerBucket.get(i)) / pStat.getNumInserts();
           insertBuckets.add(bkt);
        }
         partitionPathToInsertBuckets.put(partitionPath, insertBuckets);
      }
    }
  }

该方法相对assignUpdagtes较为复杂,其首先会获取记录中涉及到的分区,若该分区有记录插入,则需要获取该分区下所有小数据文件(小于配置的最小数据文件大小),然后计算该小文件中还可插入多少条记录,之后计算剩余待插入记录还需要多少桶,并生成INSERT类型的桶信息和随机文件名,最后遍历所有的桶,并生成InsertBucket和对应的权重信息,经过上述步骤就完成了记录插入的处理,即优先插入小文件,对于剩余的记录则写入新文件。

对于自定义Partitioner而言,最重要的是numPartitionsgetPartition两个方法,numPartitions返回桶的个数(totalBuckets),getPartition方法核心代码如下

public int getPartition(Object key) {
     Tuple2<HoodieKey, Option<HoodieRecordLocation>> keyLocation =
        (Tuple2<HoodieKey, Option<HoodieRecordLocation>>) key;
     if (keyLocation._2().isPresent()) {
       HoodieRecordLocation location = keyLocation._2().get();
       return updateLocationToBucket.get(location.getFileId());
    } else {
       List<InsertBucket> targetBuckets = partitionPathToInsertBuckets.get(keyLocation._1().getPartitionPath());
       // pick the target bucket to use based on the weights.
       double totalWeight = 0.0;
       final long totalInserts = Math.max(1, globalStat.getNumInserts());
       final long hashOfKey =
           Hashing.md5().hashString(keyLocation._1().getRecordKey(), StandardCharsets.UTF_8).asLong();
       final double r = 1.0 * Math.floorMod(hashOfKey, totalInserts) / totalInserts;
       for (InsertBucket insertBucket : targetBuckets) {
         totalWeight += insertBucket.weight;
         if (r <= totalWeight) {
           return insertBucket.bucketNumber;
        }
      }
       // return first one, by default
       return targetBuckets.get(0).bucketNumber;
    }
  }

对于更新的记录,则从文件名 -> 桶序号映射中根据文件名返回桶序号,对于插入的记录,则结合权重信息计算后返回桶序号。

再回到upsertRecordsInternal方法,定义了Partitioner后,然后对记录进行一次重新分区(一个bucket对应一个分区),然后再调用HoodieCopyOnWriteTable#handleUpsertPartition进行更新插入,其核心代码如下

public Iterator<List<WriteStatus>> handleUpsertPartition(String commitTime, Integer partition, Iterator recordItr,
     Partitioner partitioner) {
   UpsertPartitioner upsertPartitioner = (UpsertPartitioner) partitioner;
   BucketInfo binfo = upsertPartitioner.getBucketInfo(partition);
   BucketType btype = binfo.bucketType;
  ...
   if (btype.equals(BucketType.INSERT)) {
       return handleInsert(commitTime, binfo.fileIdPrefix, recordItr);
  } else if (btype.equals(BucketType.UPDATE)) {
       return handleUpdate(commitTime, binfo.fileIdPrefix, recordItr);
  } else {
       throw new HoodieUpsertException("Unknown bucketType " + btype + " for partition :" + partition);
  }
  ...
}

首先根据分区号获取到对应的桶信息,然后分别处理INSERTUPDATE不同场景,其中handleInsert方法核心代码如下

public Iterator<List<WriteStatus>> handleInsert(String commitTime, String idPfx, Iterator<HoodieRecord<T>> recordItr)
     throws Exception {
   // This is needed since sometimes some buckets are never picked in getPartition() and end up with 0 records
   if (!recordItr.hasNext()) {
     return Collections.singletonList((List<WriteStatus>) Collections.EMPTY_LIST).iterator();
  }
   return new CopyOnWriteLazyInsertIterable<>(recordItr, config, commitTime, this, idPfx);
}

如果分区无记录,则直接返回空迭代器,否则会创建一个迭代器进行处理。

handleUpdate方法核心代码如下

public Iterator<List<WriteStatus>> handleUpdate(String commitTime, String fileId, Iterator<HoodieRecord<T>> recordItr)
     throws IOException {
   // This is needed since sometimes some buckets are never picked in getPartition() and end up with 0 records
   if (!recordItr.hasNext()) {
     return Collections.singletonList((List<WriteStatus>) Collections.EMPTY_LIST).iterator();
  }
   // these are updates
   HoodieMergeHandle upsertHandle = getUpdateHandle(commitTime, fileId, recordItr);
   return handleUpdateInternal(upsertHandle, commitTime, fileId);
}
 protected Iterator<List<WriteStatus>> handleUpdateInternal(HoodieMergeHandle upsertHandle, String commitTime,
     String fileId) throws IOException {
  ...
   AvroReadSupport.setAvroReadSchema(getHadoopConf(), upsertHandle.getWriterSchema());
     BoundedInMemoryExecutor<GenericRecord, GenericRecord, Void> wrapper = null;
     try (ParquetReader<IndexedRecord> reader =
         AvroParquetReader.<IndexedRecord>builder(upsertHandle.getOldFilePath()).withConf(getHadoopConf()).build()) {
       wrapper = new SparkBoundedInMemoryExecutor(config, new ParquetReaderIterator(reader),
           new UpdateHandler(upsertHandle), x -> x);
       wrapper.execute();
  ...
   return Collections.singletonList(Collections.singletonList(upsertHandle.getWriteStatus())).iterator();
}

可以看到如果分区无记录,则直接返回空迭代器,否则会再进一步处理。

不管是对于INSERT还是UPDATE,其都会借助BoundedInMemoryExecutor来转发对记录的处理(涉及生产者-消费者-队列模型,后续会单独分析)。

对于UPDATE,最终会由HoodieMergeHandle来处理实际的更新,具体可参考

HoodieMergeHandle#write方法,其核心代码如下
 public void write(GenericRecord oldRecord) {
   String key = oldRecord.get(HoodieRecord.RECORD_KEY_METADATA_FIELD).toString();
   boolean copyOldRecord = true;
   if (keyToNewRecords.containsKey(key)) {
     // If we have duplicate records that we are updating, then the hoodie record will be deflated after
     // writing the first record. So make a copy of the record to be merged
     HoodieRecord<T> hoodieRecord = new HoodieRecord<>(keyToNewRecords.get(key));
     Option<IndexedRecord> combinedAvroRecord =
           hoodieRecord.getData().combineAndGetUpdateValue(oldRecord, useWriterSchema ? writerSchema : originalSchema);
       if (writeUpdateRecord(hoodieRecord, combinedAvroRecord)) {
         /*
          * ONLY WHEN 1) we have an update for this key AND 2) We are able to successfully write the the combined new
          * value
          *
          * We no longer need to copy the old record over.
          */
         copyOldRecord = false;
      }
       writtenRecordKeys.add(key);
  }
   if (copyOldRecord) {
    ...
     storageWriter.writeAvro(key, oldRecord);
    ...
     recordsWritten++;
  }
}

如果旧记录(文件中的旧记录)在新纪录(新写入的记录)中存在,将旧记录与新纪录合并(合并策略可以自定义实现,默认新记录覆盖旧记录),合并后再写入新文件(与原来FileId相同,但是commitTime不同,commitTime越大,文件越新),如果旧记录不存在,那么需要复制旧记录,然后写入新文件中。这样便完成了文件中已存在记录的更新和文件中未存在记录的复制,保证无记录丢失。

对于INSERT,最终会由HoodieCreateHandle来处理实际的插入,具体可参考HoodieCreateHandle#write方法,其核心代码如下

public void write(HoodieRecord record, Option<IndexedRecord> avroRecord) {
   Option recordMetadata = record.getData().getMetadata();
   if (avroRecord.isPresent()) {
       // Convert GenericRecord to GenericRecord with hoodie commit metadata in schema
       IndexedRecord recordWithMetadataInSchema = rewriteRecord((GenericRecord) avroRecord.get());
       storageWriter.writeAvroWithMetadata(recordWithMetadataInSchema, record);
       // update the new location of record, so we know where to find it next
       record.unseal();
       record.setNewLocation(new HoodieRecordLocation(instantTime, writeStatus.getFileId()));
       record.seal();
       recordsWritten++;
       insertRecordsWritten++;
    } else {
       recordsDeleted++;
    }
     writeStatus.markSuccess(record, recordMetadata);
     // deflate record payload after recording success. This will help users access payload as a
     // part of marking
     // record successful.
     record.deflate();
}

对于值存在的记录,则表示插入,写入数据文件,然后释放记录的内容,当然在调用该write方法写入之前,需要先判断该文件还能不能写入(当前文件大小是否大于配置的最大数据文件大小和分区路径是否一致),若不能写入,则会在原来FileId上加从0开始的递增序列以生成新文件写入。

经过上述步骤就完成了整个写入过程,之后还有些收尾工作,如索引的更新、写入失败处理及临时文件清除等,这里不再具体分析。

总结

对于Upsert而言,Hudi总体的处理流程是先根据索引给记录打标签,然后进行一次重新分区,对于新插入的记录,会优先插入小文件中,避免出现太多小文件,而且也会根据数据文件的具体配置控制数据文件的大小;而对于更新的记录,则会与旧记录进行合并、必要时复制旧记录到新文件(FileId与旧文件的FileId相同,commitTime不同)中。

目录
相关文章
|
7月前
|
存储 测试技术 分布式数据库
提升 Apache Hudi Upsert 性能的三个建议
提升 Apache Hudi Upsert 性能的三个建议
118 1
|
存储 SQL JSON
Delta Lake、Hudi与Iceberg详解
Delta Lake、Hudi与Iceberg详解
1032 0
Delta Lake、Hudi与Iceberg详解
|
SQL 前端开发 流计算
flink写入数据到hudi的四种方式
本文描述flink写入数据到hudi的四种方式
|
4月前
|
数据采集 Oracle 关系型数据库
实时计算 Flink版产品使用问题之怎么实现从Oracle数据库读取多个表并将数据写入到Iceberg表
实时计算Flink版作为一种强大的流处理和批处理统一的计算框架,广泛应用于各种需要实时数据处理和分析的场景。实时计算Flink版通常结合SQL接口、DataStream API、以及与上下游数据源和存储系统的丰富连接器,提供了一套全面的解决方案,以应对各种实时计算需求。其低延迟、高吞吐、容错性强的特点,使其成为众多企业和组织实时数据处理首选的技术平台。以下是实时计算Flink版的一些典型使用合集。
|
4月前
|
SQL 分布式计算 数据处理
实时计算 Flink版产品使用问题之怎么将数据从Hive表中读取并写入到另一个Hive表中
实时计算Flink版作为一种强大的流处理和批处理统一的计算框架,广泛应用于各种需要实时数据处理和分析的场景。实时计算Flink版通常结合SQL接口、DataStream API、以及与上下游数据源和存储系统的丰富连接器,提供了一套全面的解决方案,以应对各种实时计算需求。其低延迟、高吞吐、容错性强的特点,使其成为众多企业和组织实时数据处理首选的技术平台。以下是实时计算Flink版的一些典型使用合集。
|
7月前
|
消息中间件 分布式计算 Kafka
Flink 1.16.2 版本在流式读取 Iceberg upsert primary key 表方面存在一些限制
,Flink 1.16.2 版本在流式读取 Iceberg upsert primary key 表方面存在一些限制
87 2
|
7月前
|
存储 索引
Hudi MergeOnRead存储类型时Upsert分析
Hudi MergeOnRead存储类型时Upsert分析
114 0
|
7月前
|
SQL 存储 测试技术
提升50%+!Presto如何提升Hudi表查询性能?
提升50%+!Presto如何提升Hudi表查询性能?
167 0
|
存储 OLAP OLTP
Hudi、Iceberg底层索引Z-Order
Hudi、Iceberg底层索引Z-Order
304 0
|
存储 SQL 分布式计算
数据湖的选型(delta iceberg hudi)以及比对
数据湖的选型(delta iceberg hudi)以及比对
681 0