SYMBOL INDEX (4143 symbols across 300 files) FILE: benchmarks/pyspark/benchmarks/__init__.py function get_benchmark (line 38) | def get_benchmark(name: str) -> Type[Benchmark]: function list_benchmarks (line 59) | def list_benchmarks() -> List[tuple[str, str]]: FILE: benchmarks/pyspark/benchmarks/base.py class Benchmark (line 30) | class Benchmark(ABC): method __init__ (line 33) | def __init__(self, spark: SparkSession, data_path: str, mode: str): method name (line 48) | def name(cls) -> str: method description (line 54) | def description(cls) -> str: method run (line 59) | def run(self) -> Dict[str, Any]: method execute_timed (line 68) | def execute_timed(self) -> Dict[str, Any]: method _print_spark_config (line 105) | def _print_spark_config(self): method _time_operation (line 114) | def _time_operation(self, operation_fn): FILE: benchmarks/pyspark/benchmarks/shuffle.py class ShuffleBenchmark (line 32) | class ShuffleBenchmark(Benchmark): method __init__ (line 35) | def __init__(self, spark, data_path: str, mode: str, num_partitions: i... method _read_and_count (line 48) | def _read_and_count(self) -> tuple[DataFrame, int]: method _repartition (line 54) | def _repartition(self, df: DataFrame) -> DataFrame: method _write_output (line 66) | def _write_output(self, df: DataFrame, output_path: str): method run (line 70) | def run(self) -> Dict[str, Any]: class ShuffleHashBenchmark (line 101) | class ShuffleHashBenchmark(ShuffleBenchmark): method name (line 105) | def name(cls) -> str: method description (line 109) | def description(cls) -> str: method _repartition (line 112) | def _repartition(self, df: DataFrame) -> DataFrame: class ShuffleRoundRobinBenchmark (line 117) | class ShuffleRoundRobinBenchmark(ShuffleBenchmark): method name (line 121) | def name(cls) -> str: method description (line 125) | def description(cls) -> str: method _repartition (line 128) | def _repartition(self, df: DataFrame) -> DataFrame: FILE: benchmarks/pyspark/generate_data.py function generate_data (line 37) | def generate_data(output_path: str, num_rows: int, num_partitions: int): function main (line 415) | def main(): FILE: benchmarks/pyspark/run_benchmark.py function main (line 34) | def main(): FILE: benchmarks/tpc/create-iceberg-tables.py function main (line 87) | def main(benchmark: str, parquet_path: str, warehouse: str, catalog: str... FILE: benchmarks/tpc/generate-comparison.py function geomean (line 27) | def geomean(data): function get_durations (line 30) | def get_durations(result, query_key): function get_all_queries (line 37) | def get_all_queries(results): function get_common_queries (line 51) | def get_common_queries(results, labels): function check_result_consistency (line 65) | def check_result_consistency(results, labels, benchmark): function generate_query_rel_speedup_chart (line 95) | def generate_query_rel_speedup_chart(baseline, comparison, label1: str, ... function generate_query_abs_speedup_chart (line 152) | def generate_query_abs_speedup_chart(baseline, comparison, label1: str, ... function generate_query_comparison_chart (line 202) | def generate_query_comparison_chart(results, labels, benchmark: str, tit... function generate_summary (line 240) | def generate_summary(results, labels, benchmark: str, title: str, common... function query_count (line 272) | def query_count(benchmark: str): function main (line 280) | def main(files, labels, benchmark: str, title: str): FILE: benchmarks/tpc/queries/tpch/q15.sql type revenue0 (line 3) | create view revenue0 (supplier_no, total_revenue) as FILE: benchmarks/tpc/run.py function load_toml (line 43) | def load_toml(path): function _parse_toml (line 49) | def _parse_toml(text): function _parse_value (line 77) | def _parse_value(raw): function load_toml (line 98) | def load_toml(path): function resolve_env (line 144) | def resolve_env(value): function resolve_env_in_list (line 155) | def resolve_env_in_list(lst): function load_engine_config (line 159) | def load_engine_config(engine_name): function apply_env_defaults (line 175) | def apply_env_defaults(config): function apply_env_exports (line 183) | def apply_env_exports(config): function check_required_env (line 190) | def check_required_env(config): function check_common_env (line 202) | def check_common_env(): function check_benchmark_env (line 210) | def check_benchmark_env(config, benchmark): function build_spark_submit_cmd (line 233) | def build_spark_submit_cmd(config, benchmark, args): function restart_spark (line 346) | def restart_spark(): function main (line 376) | def main(): FILE: benchmarks/tpc/tpcbench.py function dedup_columns (line 36) | def dedup_columns(df): function result_hash (line 50) | def result_hash(rows): function main (line 59) | def main( FILE: common/src/main/java/org/apache/arrow/c/AbstractCometSchemaImporter.java class AbstractCometSchemaImporter (line 29) | public abstract class AbstractCometSchemaImporter { method AbstractCometSchemaImporter (line 34) | public AbstractCometSchemaImporter(BufferAllocator allocator) { method getAllocator (line 39) | public BufferAllocator getAllocator() { method getProvider (line 43) | public CDataDictionaryProvider getProvider() { method importField (line 47) | public Field importField(ArrowSchema schema) { method importVector (line 64) | public FieldVector importVector(ArrowArray array, ArrowSchema schema) { method close (line 72) | @IcebergApi FILE: common/src/main/java/org/apache/arrow/c/ArrowImporter.java class ArrowImporter (line 33) | public class ArrowImporter { method ArrowImporter (line 37) | public ArrowImporter(BufferAllocator allocator) { method importField (line 42) | Field importField(ArrowSchema schema, CDataDictionaryProvider provider) { method importVector (line 51) | public FieldVector importVector( FILE: common/src/main/java/org/apache/comet/CometNativeException.java class CometNativeException (line 23) | public class CometNativeException extends CometRuntimeException { method CometNativeException (line 24) | public CometNativeException(String message) { FILE: common/src/main/java/org/apache/comet/CometOutOfMemoryError.java class CometOutOfMemoryError (line 23) | public class CometOutOfMemoryError extends OutOfMemoryError { method CometOutOfMemoryError (line 24) | public CometOutOfMemoryError(String msg) { FILE: common/src/main/java/org/apache/comet/CometRuntimeException.java class CometRuntimeException (line 23) | public class CometRuntimeException extends RuntimeException { method CometRuntimeException (line 24) | public CometRuntimeException(String message) { method CometRuntimeException (line 28) | public CometRuntimeException(String message, Throwable cause) { FILE: common/src/main/java/org/apache/comet/CometSchemaImporter.java class CometSchemaImporter (line 26) | @IcebergApi method CometSchemaImporter (line 28) | @IcebergApi FILE: common/src/main/java/org/apache/comet/NativeBase.java class NativeBase (line 41) | public abstract class NativeBase { method isLoaded (line 64) | public static synchronized boolean isLoaded() throws Throwable { method setLoaded (line 72) | static synchronized void setLoaded(boolean b) { method load (line 76) | static synchronized void load() { method bundleLoadLibrary (line 111) | private static void bundleLoadLibrary() { method initWithLogConf (line 155) | private static void initWithLogConf() { method cleanupOldTempLibs (line 178) | private static void cleanupOldTempLibs() { method setArrowProperties (line 206) | private static void setArrowProperties() { method setPropertyIfNull (line 211) | private static void setPropertyIfNull(String key, String value) { type OS (line 224) | private enum OS { method OS (line 232) | OS(String name, String libExtension) { method arch (line 238) | private static String arch() { method os (line 242) | private static OS os() { method checkArch (line 260) | private static boolean checkArch() { method resourceName (line 282) | private static String resourceName() { method init (line 294) | static native void init(String logConfPath, String logLevel); method isFeatureEnabled (line 302) | public static native boolean isFeatureEnabled(String featureName); FILE: common/src/main/java/org/apache/comet/ParquetRuntimeException.java class ParquetRuntimeException (line 23) | public class ParquetRuntimeException extends CometRuntimeException { method ParquetRuntimeException (line 24) | public ParquetRuntimeException(String message) { method ParquetRuntimeException (line 28) | public ParquetRuntimeException(String message, Throwable cause) { FILE: common/src/main/java/org/apache/comet/exceptions/CometQueryExecutionException.java class CometQueryExecutionException (line 41) | public final class CometQueryExecutionException extends CometNativeExcep... method CometQueryExecutionException (line 48) | public CometQueryExecutionException(String jsonMessage) { method isJsonMessage (line 58) | public boolean isJsonMessage() { FILE: common/src/main/java/org/apache/comet/parquet/AbstractColumnReader.java class AbstractColumnReader (line 35) | @IcebergApi method AbstractColumnReader (line 68) | AbstractColumnReader( method AbstractColumnReader (line 81) | AbstractColumnReader( method getDescriptor (line 90) | ColumnDescriptor getDescriptor() { method getPath (line 94) | String getPath() { method setBatchSize (line 101) | @IcebergApi method readBatch (line 114) | public abstract void readBatch(int total); method currentBatch (line 117) | public abstract CometVector currentBatch(); method close (line 119) | @IcebergApi method initNative (line 129) | protected void initNative() { FILE: common/src/main/java/org/apache/comet/parquet/ArrowConstantColumnReader.java class ArrowConstantColumnReader (line 45) | public class ArrowConstantColumnReader extends AbstractColumnReader { method ArrowConstantColumnReader (line 55) | ArrowConstantColumnReader(StructField field, int batchSize, boolean us... method ArrowConstantColumnReader (line 65) | ArrowConstantColumnReader( method setBatchSize (line 74) | @Override method readBatch (line 81) | @Override method currentBatch (line 89) | @Override method close (line 94) | @Override method initVector (line 106) | private void initVector(Object value, int count) { method createNullVector (line 119) | private FieldVector createNullVector(int count) { method createFilledVector (line 159) | private FieldVector createFilledVector(Object value, int count) { FILE: common/src/main/java/org/apache/comet/parquet/ArrowRowIndexColumnReader.java class ArrowRowIndexColumnReader (line 38) | public class ArrowRowIndexColumnReader extends AbstractColumnReader { method ArrowRowIndexColumnReader (line 50) | public ArrowRowIndexColumnReader(StructField field, int batchSize, lon... method setBatchSize (line 56) | @Override method readBatch (line 62) | @Override method currentBatch (line 93) | @Override method close (line 98) | @Override FILE: common/src/main/java/org/apache/comet/parquet/BloomFilterReader.java class BloomFilterReader (line 47) | public class BloomFilterReader implements FilterPredicate.Visitor, U extends UserDefinedPredicate> B... method readBloomFilter (line 159) | BloomFilter readBloomFilter(ColumnChunkMetaData meta) { method readBloomFilterInternal (line 180) | private BloomFilter readBloomFilterInternal(ColumnChunkMetaData meta) ... FILE: common/src/main/java/org/apache/comet/parquet/ColumnIndexReader.java class ColumnIndexReader (line 47) | class ColumnIndexReader implements ColumnIndexStore { method getColumnIndex (line 53) | @Override method getOffsetIndex (line 58) | @Override method getColumnIndex (line 66) | @Override method getOffsetIndex (line 71) | @Override method create (line 85) | static ColumnIndexReader create( method ColumnIndexReader (line 97) | private ColumnIndexReader( method getColumnIndex (line 114) | @Override method getOffsetIndex (line 119) | @Override type IndexStore (line 124) | private interface IndexStore { method getColumnIndex (line 125) | ColumnIndex getColumnIndex(); method getOffsetIndex (line 127) | OffsetIndex getOffsetIndex(); class IndexStoreImpl (line 130) | private class IndexStoreImpl implements IndexStore { method IndexStoreImpl (line 136) | IndexStoreImpl(ColumnChunkMetaData meta) { method getColumnIndex (line 153) | @Override method getOffsetIndex (line 168) | @Override method readColumnIndex (line 175) | ColumnIndex readColumnIndex(ColumnChunkMetaData column) throws IOExcep... method readOffsetIndex (line 204) | OffsetIndex readOffsetIndex(ColumnChunkMetaData column) throws IOExcep... FILE: common/src/main/java/org/apache/comet/parquet/ColumnPageReader.java class ColumnPageReader (line 40) | public class ColumnPageReader implements PageReader { method ColumnPageReader (line 54) | ColumnPageReader( method getTotalValueCount (line 93) | @Override method getPageValueCount (line 99) | public int getPageValueCount() { method skipPage (line 104) | public void skipPage() { method readPage (line 109) | @Override method readDictionaryPage (line 225) | @Override method getPageOrdinal (line 249) | private int getPageOrdinal(int currentPageIndex) { FILE: common/src/main/java/org/apache/comet/parquet/ColumnReader.java class ColumnReader (line 52) | @IcebergApi method ColumnReader (line 96) | ColumnReader( method setPageReader (line 116) | @IcebergApi method setRowGroupReader (line 132) | @IcebergApi method readBatch (line 139) | @Override method currentBatch (line 169) | @Override method close (line 174) | @Override method loadVector (line 184) | public CometDecodedVector loadVector() { method readPage (line 250) | protected void readPage() { method isValidValueEncoding (line 303) | @SuppressWarnings("deprecation") FILE: common/src/main/java/org/apache/comet/parquet/CometFileKeyUnwrapper.java class CometFileKeyUnwrapper (line 93) | public class CometFileKeyUnwrapper { method normalizeS3Scheme (line 112) | private String normalizeS3Scheme(final String filePath) { method storeDecryptionKeyRetriever (line 131) | public void storeDecryptionKeyRetriever(final String filePath, final C... method getKey (line 159) | public byte[] getKey(final String filePath, final byte[] keyMetadata) FILE: common/src/main/java/org/apache/comet/parquet/CometInputFile.java class CometInputFile (line 42) | public class CometInputFile implements InputFile { method fromPath (line 50) | public static CometInputFile fromPath(Path path, Configuration conf) t... method CometInputFile (line 55) | private CometInputFile(FileSystem fs, FileStatus stat, Configuration c... method getLength (line 61) | @Override method getConf (line 66) | public Configuration getConf() { method getFileSystem (line 70) | public FileSystem getFileSystem() { method getPath (line 74) | public Path getPath() { method newStream (line 78) | @Override method newStream (line 105) | public SeekableInputStream newStream(long offset, long length) throws ... method toString (line 137) | @Override method isAtLeastHadoop33 (line 142) | private static boolean isAtLeastHadoop33() { method isAtLeastHadoop33 (line 147) | static boolean isAtLeastHadoop33(String version) { FILE: common/src/main/java/org/apache/comet/parquet/DictionaryPageReader.java class DictionaryPageReader (line 47) | public class DictionaryPageReader implements DictionaryPageReadStore { method DictionaryPageReader (line 54) | DictionaryPageReader( method readDictionaryPage (line 70) | @Override method readDictionary (line 98) | DictionaryPage readDictionary(ColumnChunkMetaData meta) throws IOExcep... method readCompressedDictionary (line 157) | private DictionaryPage readCompressedDictionary( method reusableCopy (line 184) | private static DictionaryPage reusableCopy(DictionaryPage dict) throws... FILE: common/src/main/java/org/apache/comet/parquet/FileReader.java class FileReader (line 106) | @IcebergApi method FileReader (line 135) | FileReader(InputFile file, ParquetReadOptions options, ReadOptions com... method FileReader (line 141) | @IcebergApi method FileReader (line 183) | FileReader( method FileReader (line 192) | FileReader( method getFooter (line 232) | ParquetMetadata getFooter() { method getFileMetaData (line 237) | FileMetaData getFileMetaData() { method getInputStream (line 242) | public SeekableInputStream getInputStream() { method getOptions (line 247) | public ParquetReadOptions getOptions() { method getRowGroups (line 252) | public List getRowGroups() { method setRequestedSchema (line 257) | public void setRequestedSchema(List projection) { method setRequestedSchemaFromSpecs (line 265) | @IcebergApi method buildParquetReadOptions (line 274) | private static ParquetReadOptions buildParquetReadOptions( method getRecordCount (line 319) | public long getRecordCount() { method getFilteredRecordCount (line 331) | public long getFilteredRecordCount() { method skipNextRowGroup (line 344) | @IcebergApi method readNextRowGroup (line 353) | @IcebergApi method readNextFilteredRowGroup (line 404) | public PageReadStore readNextFilteredRowGroup() throws IOException { method getColumnIndexReader (line 462) | ColumnIndexReader getColumnIndexReader(int blockIndex) { method readChunks (line 471) | private RowGroupReader readChunks( method shouldReadParallel (line 490) | private boolean shouldReadParallel() { method shouldReadParallel (line 499) | static boolean shouldReadParallel(ReadOptions options, String scheme) { method shouldReadParallelForScheme (line 503) | private static boolean shouldReadParallelForScheme(String scheme) { class ReadRange (line 517) | static class ReadRange { method toString (line 523) | @Override method getReadRanges (line 536) | List getReadRanges(List allParts, int ... method readAllRangesParallel (line 569) | private void readAllRangesParallel(List allRanges) { method readAllPartsParallel (line 627) | public void readAllPartsParallel(List allParts, C... method readChunkPages (line 656) | private void readChunkPages(Chunk chunk, BlockMetaData block) throws I... method advanceToNextBlock (line 678) | private boolean advanceToNextBlock() { method getRowIndices (line 687) | public long[] getRowIndices() { method getRowIndices (line 691) | public static long[] getRowIndices(List blocks) { method getRowIndexOffset (line 705) | public static long getRowIndexOffset(BlockMetaData metaData) { method getRowRanges (line 715) | private RowRanges getRowRanges(int blockIndex) { method readFooter (line 732) | private static ParquetMetadata readFooter( method filterRowGroups (line 812) | private List filterRowGroups(List blocks) { method filterRowGroups (line 816) | public static List filterRowGroups( method filterRowGroups (line 840) | public static List filterRowGroups( method listWithNulls (line 864) | private static List listWithNulls(int size) { method closeStream (line 868) | public void closeStream() throws IOException { method close (line 874) | @IcebergApi class ChunkListBuilder (line 891) | private class ChunkListBuilder { class ChunkData (line 892) | private class ChunkData { method add (line 899) | void add(ChunkDescriptor descriptor, List buffers) { method setOffsetIndex (line 908) | void setOffsetIndex(ChunkDescriptor descriptor, OffsetIndex offsetIn... method build (line 917) | List build() { class Chunk (line 929) | private class Chunk { method Chunk (line 939) | Chunk(ChunkDescriptor descriptor, List buffers, OffsetIn... method readPageHeader (line 945) | protected PageHeader readPageHeader(BlockCipher.Decryptor blockDecry... method verifyCrc (line 954) | private void verifyCrc(int referenceCrc, byte[] bytes, String except... method readAllPages (line 962) | private ColumnPageReader readAllPages() throws IOException { method readAllPages (line 966) | private ColumnPageReader readAllPages( method hasMorePages (line 1128) | private boolean hasMorePages(long valuesCountReadSoFar, int dataPage... method getPageOrdinal (line 1134) | private int getPageOrdinal(int dataPageCountReadSoFar) { method readAsBytesInput (line 1147) | public BytesInput readAsBytesInput(int size) throws IOException { class ConsecutivePartList (line 1156) | private class ConsecutivePartList { method ConsecutivePartList (line 1170) | ConsecutivePartList(long offset) { method addChunk (line 1188) | public void addChunk(ChunkDescriptor descriptor) { method allocateReadBuffers (line 1193) | private void allocateReadBuffers() { method readAll (line 1214) | public void readAll(SeekableInputStream f, ChunkListBuilder builder)... method setReadMetrics (line 1239) | private void setReadMetrics(long startNs) { method endPos (line 1265) | public long endPos() { class ChunkDescriptor (line 1271) | private static class ChunkDescriptor { method ChunkDescriptor (line 1284) | ChunkDescriptor( method hashCode (line 1292) | @Override method equals (line 1297) | @Override FILE: common/src/main/java/org/apache/comet/parquet/FooterReader.java class FooterReader (line 39) | public class FooterReader { method readFooter (line 40) | public static ParquetMetadata readFooter(Configuration configuration, ... FILE: common/src/main/java/org/apache/comet/parquet/IcebergCometNativeBatchReader.java class IcebergCometNativeBatchReader (line 34) | public class IcebergCometNativeBatchReader extends NativeBatchReader { method IcebergCometNativeBatchReader (line 36) | public IcebergCometNativeBatchReader(StructType requiredSchema) { method init (line 42) | public void init( method getSparkSchema (line 82) | public StructType getSparkSchema() { FILE: common/src/main/java/org/apache/comet/parquet/IndexFilter.java class IndexFilter (line 29) | public class IndexFilter { method IndexFilter (line 34) | public IndexFilter(RowRanges rowRanges, OffsetIndex offsetIndex, long ... method filterOffsetIndex (line 40) | OffsetIndex filterOffsetIndex() { method calculateOffsetRanges (line 56) | List calculateOffsetRanges(OffsetIndex filteredOffsetInde... class FilteredOffsetIndex (line 83) | private static class FilteredOffsetIndex implements OffsetIndex { method FilteredOffsetIndex (line 87) | private FilteredOffsetIndex(OffsetIndex offsetIndex, int[] indexMap) { method getPageOrdinal (line 92) | @Override method getPageCount (line 97) | @Override method getOffset (line 102) | @Override method getCompressedPageSize (line 107) | @Override method getFirstRowIndex (line 112) | @Override method getLastRowIndex (line 117) | @Override class OffsetRange (line 127) | static class OffsetRange { method OffsetRange (line 131) | private OffsetRange(long offset, int length) { method extend (line 136) | private boolean extend(long offset, int length) { FILE: common/src/main/java/org/apache/comet/parquet/LazyColumnReader.java class LazyColumnReader (line 32) | public class LazyColumnReader extends ColumnReader { method LazyColumnReader (line 46) | LazyColumnReader( method setPageReader (line 58) | @Override method readBatch (line 74) | @Override method currentBatch (line 90) | @Override method readAllBatch (line 96) | public void readAllBatch() { method materializeUpToIfNecessary (line 110) | public boolean materializeUpToIfNecessary(int rowId) { method readBatch (line 129) | private void readBatch(int rowId, int numNullRowsToPad) { method tryPageSkip (line 160) | private int tryPageSkip(int rowId) { FILE: common/src/main/java/org/apache/comet/parquet/Native.java class Native (line 27) | public final class Native extends NativeBase { method readBatch (line 28) | public static int[] readBatch(long handle, int batchSize) { method skipBatch (line 32) | public static int skipBatch(long handle, int batchSize) { method initColumnReader (line 66) | public static native long initColumnReader( method setDictionaryPage (line 97) | public static native void setDictionaryPage( method setPageV1 (line 108) | public static native void setPageV1( method setPageV2 (line 121) | public static native void setPageV2( method resetBatch (line 135) | @IcebergApi method readBatch (line 160) | public static native int[] readBatch(long handle, int batchSize, int n... method skipBatch (line 180) | public static native int skipBatch(long handle, int batchSize, boolean... method currentBatch (line 189) | public static native void currentBatch(long handle, long arrayAddr, lo... method closeColumnReader (line 196) | public static native void closeColumnReader(long handle); method validateObjectStoreConfig (line 207) | public static native void validateObjectStoreConfig( method initRecordBatchReader (line 218) | public static native long initRecordBatchReader( method readNextRecordBatch (line 241) | public static native int readNextRecordBatch(long handle); method currentColumnBatch (line 254) | public static native void currentColumnBatch( method closeRecordBatchReader (line 264) | public static native void closeRecordBatchReader(long handle); FILE: common/src/main/java/org/apache/comet/parquet/NativeBatchReader.java class NativeBatchReader (line 104) | public class NativeBatchReader extends RecordReader... class FileInfo (line 110) | public static class FileInfo { method FileInfo (line 116) | public FileInfo(long start, long length, String filePath, long fileS... method fromPartitionedFile (line 128) | public static FileInfo fromPartitionedFile(PartitionedFile file) thr... method start (line 132) | public long start() { method length (line 136) | public long length() { method filePath (line 140) | public String filePath() { method fileSize (line 144) | public long fileSize() { method pathUri (line 148) | public URI pathUri() throws URISyntaxException { method NativeBatchReader (line 218) | protected NativeBatchReader() { method NativeBatchReader (line 224) | public NativeBatchReader(String file, int capacity) { method NativeBatchReader (line 229) | public NativeBatchReader( method NativeBatchReader (line 235) | public NativeBatchReader( method NativeBatchReader (line 256) | private NativeBatchReader(AbstractColumnReader[] columnReaders) { method NativeBatchReader (line 269) | NativeBatchReader( method NativeBatchReader (line 304) | NativeBatchReader( method init (line 342) | public void init() throws Throwable { method getParquetColumn (line 597) | private ParquetColumn getParquetColumn(MessageType schema, StructType ... method getIdToParquetFieldMap (line 610) | private Map> getIdToParquetFieldMap(GroupType type) { method getCaseSensitiveParquetFieldMap (line 616) | private Map> getCaseSensitiveParquetFieldMap(GroupT... method getCaseInsensitiveParquetFieldMap (line 620) | private Map> getCaseInsensitiveParquetFieldMap(Grou... method buildCaseSensitiveNameMap (line 625) | private Map buildCaseSensitiveNameMap(List types) { method buildCaseInsensitiveNameMap (line 629) | private Map buildCaseInsensitiveNameMap(List types) { method getMatchingParquetFieldById (line 634) | private Type getMatchingParquetFieldById( method getMatchingNameById (line 664) | private String getMatchingNameById( method getSparkSchemaByFieldId (line 682) | private StructType getSparkSchemaByFieldId( method isPrimitiveCatalystType (line 714) | private static boolean isPrimitiveCatalystType(DataType dataType) { method getSparkTypeByFieldId (line 720) | private DataType getSparkTypeByFieldId( method getSparkArrayTypeByFieldId (line 762) | private DataType getSparkArrayTypeByFieldId( method checkParquetType (line 793) | private void checkParquetType(ParquetColumn column) throws IOException { method getColumnIndexFromParquetColumn (line 825) | private int getColumnIndexFromParquetColumn(ParquetColumn column) { method checkColumn (line 849) | private void checkColumn(ParquetColumn column) throws IOException { method containsPath (line 877) | private boolean containsPath(Type parquetType, String[] path) { method containsPath (line 881) | private boolean containsPath(Type parquetType, String[] path, int dept... method setSparkSchema (line 893) | public void setSparkSchema(StructType schema) { method getColumnReaders (line 897) | public AbstractColumnReader[] getColumnReaders() { method initialize (line 901) | @Override method nextKeyValue (line 907) | @Override method getCurrentKey (line 912) | @Override method getCurrentValue (line 917) | @Override method getProgress (line 922) | @Override method currentBatch (line 932) | public ColumnarBatch currentBatch() { method nextBatch (line 941) | public boolean nextBatch() throws IOException { method close (line 987) | @Override method loadNextBatch (line 1007) | @SuppressWarnings("deprecation") method getTaskAccumulator (line 1067) | @SuppressWarnings("unchecked") method serializeArrowSchema (line 1087) | private byte[] serializeArrowSchema(Schema schema) throws IOException { FILE: common/src/main/java/org/apache/comet/parquet/NativeColumnReader.java class NativeColumnReader (line 39) | public class NativeColumnReader extends AbstractColumnReader { method NativeColumnReader (line 74) | NativeColumnReader( method initNative (line 95) | @Override method readBatch (line 103) | @Override method currentBatch (line 111) | @Override method close (line 116) | @Override method loadVector (line 126) | public CometDecodedVector loadVector() { FILE: common/src/main/java/org/apache/comet/parquet/ParquetColumnSpec.java class ParquetColumnSpec (line 32) | @IcebergApi method ParquetColumnSpec (line 47) | @IcebergApi method getFieldId (line 69) | @IcebergApi method getPath (line 74) | @IcebergApi method getPhysicalType (line 79) | @IcebergApi method getTypeLength (line 84) | @IcebergApi method isRepeated (line 89) | public boolean isRepeated() { method getMaxRepetitionLevel (line 93) | @IcebergApi method getMaxDefinitionLevel (line 98) | @IcebergApi method getLogicalTypeName (line 103) | @IcebergApi method getLogicalTypeParams (line 108) | @IcebergApi FILE: common/src/main/java/org/apache/comet/parquet/ParquetMetadataSerializer.java class ParquetMetadataSerializer (line 36) | public class ParquetMetadataSerializer { method ParquetMetadataSerializer (line 40) | public ParquetMetadataSerializer() { method ParquetMetadataSerializer (line 44) | public ParquetMetadataSerializer(ParquetMetadataConverter converter) { method serialize (line 55) | public byte[] serialize(ParquetMetadata metadata) throws IOException { method deserialize (line 69) | public ParquetMetadata deserialize(byte[] bytes) throws IOException { FILE: common/src/main/java/org/apache/comet/parquet/ReadOptions.java class ReadOptions (line 37) | @IcebergApi method ReadOptions (line 58) | ReadOptions( method isParallelIOEnabled (line 71) | public boolean isParallelIOEnabled() { method parallelIOThreadPoolSize (line 75) | public int parallelIOThreadPoolSize() { method isIOMergeRangesEnabled (line 79) | public boolean isIOMergeRangesEnabled() { method getIOMergeRangesDelta (line 83) | public int getIOMergeRangesDelta() { method adjustReadRangesSkew (line 87) | public boolean adjustReadRangesSkew() { method builder (line 91) | @IcebergApi class Builder (line 96) | @IcebergApi method enableParallelIO (line 110) | public Builder enableParallelIO(boolean b) { method withParallelIOThreadPoolSize (line 121) | public Builder withParallelIOThreadPoolSize(int numThreads) { method enableIOMergeRanges (line 126) | public Builder enableIOMergeRanges(boolean enableIOMergeRanges) { method withIOMergeRangesDelta (line 131) | public Builder withIOMergeRangesDelta(int ioMergeRangesDelta) { method adjustReadRangeSkew (line 136) | public Builder adjustReadRangeSkew(boolean adjustReadRangeSkew) { method build (line 141) | @IcebergApi method Builder (line 151) | @IcebergApi method setS3Config (line 180) | private void setS3Config() { method setS3ConfIfGreater (line 197) | private void setS3ConfIfGreater(Configuration conf, String key, int ... method setS3ConfIfGreater (line 209) | private void setS3ConfIfGreater(Configuration conf, String key, long... FILE: common/src/main/java/org/apache/comet/parquet/RowGroupFilter.java class RowGroupFilter (line 36) | public class RowGroupFilter implements Visitor> { type FilterLevel (line 42) | public enum FilterLevel { method filterRowGroups (line 48) | public static List filterRowGroups( method filterRowGroups (line 53) | public static List filterRowGroups( method RowGroupFilter (line 58) | private RowGroupFilter(List levels, List b... method RowGroupFilter (line 65) | private RowGroupFilter(List levels, List b... method visit (line 72) | @Override method visit (line 115) | @Override method visit (line 121) | @Override FILE: common/src/main/java/org/apache/comet/parquet/RowGroupReader.java class RowGroupReader (line 34) | @IcebergApi method RowGroupReader (line 41) | public RowGroupReader(long rowCount, long rowIndexOffset) { method RowGroupReader (line 47) | RowGroupReader(RowRanges rowRanges) { method getRowCount (line 53) | @IcebergApi method getPageReader (line 59) | @Override method getPageReader (line 64) | public PageReader getPageReader(String[] path) { method getRowIndexes (line 73) | @Override method getRowIndexOffset (line 78) | @Override method addColumn (line 83) | void addColumn(ColumnDescriptor path, ColumnPageReader reader) { FILE: common/src/main/java/org/apache/comet/parquet/TypeUtil.java class TypeUtil (line 37) | public class TypeUtil { method convertToParquet (line 44) | @IcebergApi method convertToParquetSpec (line 115) | public static ParquetColumnSpec convertToParquetSpec(StructField field) { method checkParquetType (line 129) | public static void checkParquetType(ColumnDescriptor descriptor, DataT... method validateTimestampType (line 244) | private static void validateTimestampType( method convertErrorForTimestampNTZ (line 256) | private static void convertErrorForTimestampNTZ(String parquetType) { method canReadAsIntDecimal (line 264) | private static boolean canReadAsIntDecimal(ColumnDescriptor descriptor... method canReadAsLongDecimal (line 270) | private static boolean canReadAsLongDecimal(ColumnDescriptor descripto... method canReadAsBinaryDecimal (line 276) | private static boolean canReadAsBinaryDecimal(ColumnDescriptor descrip... method isLongDecimal (line 281) | private static boolean isLongDecimal(DataType dt) { method isDecimalTypeMatched (line 289) | private static boolean isDecimalTypeMatched(ColumnDescriptor descripto... method isTimestampTypeMatched (line 323) | private static boolean isTimestampTypeMatched( method isUnsignedIntTypeMatched (line 329) | private static boolean isUnsignedIntTypeMatched( method isSpark40Plus (line 336) | static boolean isSpark40Plus() { FILE: common/src/main/java/org/apache/comet/parquet/Utils.java class Utils (line 35) | public class Utils { method getColumnReader (line 38) | @IcebergApi method getColumnReader (line 64) | @IcebergApi method getColumnReader (line 77) | public static ColumnReader getColumnReader( method supportLazyMaterialization (line 94) | private static boolean supportLazyMaterialization(DataType type) { method initColumnReader (line 110) | public static long initColumnReader( class TypePromotionInfo (line 179) | static class TypePromotionInfo { method TypePromotionInfo (line 190) | TypePromotionInfo(int physicalTypeId, int precision, int scale, int ... method TypePromotionInfo (line 197) | TypePromotionInfo(DataType sparkReadType) { method getPhysicalTypeId (line 233) | static int getPhysicalTypeId(PrimitiveType.PrimitiveTypeName typeName) { method getLogicalTypeId (line 262) | static int getLogicalTypeId(LogicalTypeAnnotation annotation) { method getTimeUnitId (line 284) | static int getTimeUnitId(LogicalTypeAnnotation.TimeUnit tu) { method buildColumnDescriptor (line 297) | @IcebergApi method reconstructLogicalType (line 343) | private static LogicalTypeAnnotation reconstructLogicalType( method descriptorToParquetColumnSpec (line 463) | @IcebergApi FILE: common/src/main/java/org/apache/comet/parquet/WrappedInputFile.java class WrappedInputFile (line 35) | @IcebergApi method WrappedInputFile (line 39) | @IcebergApi method getLength (line 44) | @Override method newStream (line 55) | @Override method toString (line 67) | @Override FILE: common/src/main/java/org/apache/comet/parquet/WrappedSeekableInputStream.java class WrappedSeekableInputStream (line 34) | public class WrappedSeekableInputStream extends DelegatingSeekableInputS... method WrappedSeekableInputStream (line 38) | public WrappedSeekableInputStream(InputStream inputStream) { method getPos (line 43) | @Override method seek (line 54) | @Override FILE: common/src/main/java/org/apache/comet/vector/CometDecodedVector.java class CometDecodedVector (line 29) | public abstract class CometDecodedVector extends CometVector { method CometDecodedVector (line 43) | protected CometDecodedVector(ValueVector vector, Field valueField, boo... method CometDecodedVector (line 47) | protected CometDecodedVector( method getValueVector (line 57) | @Override method setNumNulls (line 62) | @Override method setNumValues (line 71) | @Override method numValues (line 87) | public int numValues() { method hasNull (line 91) | @Override method numNulls (line 96) | @Override method isNullAt (line 101) | @Override FILE: common/src/main/java/org/apache/comet/vector/CometDelegateVector.java class CometDelegateVector (line 32) | public class CometDelegateVector extends CometVector { method CometDelegateVector (line 35) | public CometDelegateVector(DataType dataType) { method CometDelegateVector (line 39) | public CometDelegateVector(DataType dataType, boolean useDecimal128) { method CometDelegateVector (line 43) | public CometDelegateVector(DataType dataType, CometVector delegate, bo... method setDelegate (line 51) | protected void setDelegate(CometVector delegate) { method setNumNulls (line 55) | @Override method setNumValues (line 60) | @Override method numValues (line 65) | @Override method hasNull (line 70) | @Override method numNulls (line 75) | @Override method isNullAt (line 80) | @Override method getBoolean (line 85) | @Override method getByte (line 90) | @Override method getShort (line 95) | @Override method getInt (line 100) | @Override method getLong (line 105) | @Override method getLongDecimal (line 110) | @Override method getFloat (line 115) | @Override method getDouble (line 120) | @Override method getDecimal (line 125) | @Override method getBinaryDecimal (line 130) | @Override method getUTF8String (line 135) | @Override method getBinary (line 140) | @Override method getArray (line 145) | @Override method getMap (line 150) | @Override method getChild (line 155) | @Override method getValueVector (line 160) | @Override method slice (line 165) | @Override method getDictionaryProvider (line 170) | @Override FILE: common/src/main/java/org/apache/comet/vector/CometDictionary.java class CometDictionary (line 26) | public class CometDictionary implements AutoCloseable { method CometDictionary (line 35) | public CometDictionary(CometPlainVector values) { method setDictionaryVector (line 40) | public void setDictionaryVector(CometPlainVector values) { method getValueVector (line 47) | public ValueVector getValueVector() { method decodeToBoolean (line 51) | public boolean decodeToBoolean(int index) { method decodeToByte (line 55) | public byte decodeToByte(int index) { method decodeToShort (line 59) | public short decodeToShort(int index) { method decodeToInt (line 63) | public int decodeToInt(int index) { method decodeToLong (line 67) | public long decodeToLong(int index) { method decodeToLongDecimal (line 71) | public long decodeToLongDecimal(int index) { method decodeToFloat (line 75) | public float decodeToFloat(int index) { method decodeToDouble (line 79) | public double decodeToDouble(int index) { method decodeToBinary (line 83) | public byte[] decodeToBinary(int index) { method decodeToUTF8String (line 109) | public UTF8String decodeToUTF8String(int index) { method close (line 113) | @Override class ByteArrayWrapper (line 118) | private static class ByteArrayWrapper { method ByteArrayWrapper (line 121) | ByteArrayWrapper(byte[] bytes) { FILE: common/src/main/java/org/apache/comet/vector/CometDictionaryVector.java class CometDictionaryVector (line 29) | public class CometDictionaryVector extends CometDecodedVector { method CometDictionaryVector (line 37) | public CometDictionaryVector( method CometDictionaryVector (line 45) | public CometDictionaryVector( method getDictionaryProvider (line 61) | @Override method close (line 66) | @Override method getBoolean (line 75) | @Override method getByte (line 80) | @Override method getShort (line 85) | @Override method getInt (line 90) | @Override method getLong (line 95) | @Override method getLongDecimal (line 100) | @Override method getFloat (line 105) | @Override method getDouble (line 110) | @Override method getUTF8String (line 115) | @Override method getBinary (line 120) | @Override method getBinaryDecimal (line 125) | @Override method slice (line 130) | @Override FILE: common/src/main/java/org/apache/comet/vector/CometLazyVector.java class CometLazyVector (line 27) | public class CometLazyVector extends CometDelegateVector { method CometLazyVector (line 30) | public CometLazyVector(DataType type, LazyColumnReader columnReader, b... method getDecodedVector (line 35) | public CometDecodedVector getDecodedVector() { method getValueVector (line 39) | @Override method setNumNulls (line 46) | @Override method setNumValues (line 51) | @Override method close (line 56) | @Override method hasNull (line 61) | @Override method numNulls (line 68) | @Override method isNullAt (line 75) | @Override FILE: common/src/main/java/org/apache/comet/vector/CometListVector.java class CometListVector (line 30) | public class CometListVector extends CometDecodedVector { method CometListVector (line 36) | public CometListVector( method getArray (line 46) | @Override method slice (line 55) | @Override FILE: common/src/main/java/org/apache/comet/vector/CometMapVector.java class CometMapVector (line 31) | public class CometMapVector extends CometDecodedVector { method CometMapVector (line 40) | public CometMapVector( method getMap (line 66) | @Override method slice (line 75) | @Override FILE: common/src/main/java/org/apache/comet/vector/CometPlainVector.java class CometPlainVector (line 34) | public class CometPlainVector extends CometDecodedVector { method CometPlainVector (line 43) | public CometPlainVector(ValueVector vector, boolean useDecimal128) { method CometPlainVector (line 47) | public CometPlainVector(ValueVector vector, boolean useDecimal128, boo... method CometPlainVector (line 51) | public CometPlainVector( method isReused (line 65) | public boolean isReused() { method setReused (line 69) | public void setReused(boolean isReused) { method setNumNulls (line 73) | @Override method getBoolean (line 79) | @Override method getByte (line 89) | @Override method getShort (line 94) | @Override method getInt (line 99) | @Override method getLong (line 104) | @Override method getLongDecimal (line 109) | @Override method getFloat (line 114) | @Override method getDouble (line 119) | @Override method getUTF8String (line 124) | @Override method getBinary (line 149) | @Override method getDictionaryProvider (line 172) | @Override method isNullAt (line 177) | @Override method slice (line 182) | @Override method convertToUuid (line 190) | private static UUID convertToUuid(byte[] buf) { FILE: common/src/main/java/org/apache/comet/vector/CometSelectionVector.java class CometSelectionVector (line 42) | public class CometSelectionVector extends CometVector { method CometSelectionVector (line 73) | public CometSelectionVector(CometVector values, int[] indices, int num... method getValuesIndex (line 111) | private int getValuesIndex(int selectionIndex) { method getValues (line 126) | public CometVector getValues() { method getIndices (line 135) | public CometVector getIndices() { method getSelectedIndices (line 144) | private int[] getSelectedIndices() { method numValues (line 148) | @Override method setNumValues (line 153) | @Override method setNumNulls (line 160) | @Override method hasNull (line 167) | @Override method numNulls (line 172) | @Override method isNullAt (line 178) | @Override method getBoolean (line 183) | @Override method getByte (line 188) | @Override method getShort (line 193) | @Override method getInt (line 198) | @Override method getLong (line 203) | @Override method getLongDecimal (line 208) | @Override method getFloat (line 213) | @Override method getDouble (line 218) | @Override method getUTF8String (line 223) | @Override method getBinary (line 228) | @Override method getArray (line 233) | @Override method getMap (line 238) | @Override method getChild (line 243) | @Override method getDictionaryProvider (line 249) | @Override method slice (line 254) | @Override method getValueVector (line 268) | @Override method close (line 273) | @Override FILE: common/src/main/java/org/apache/comet/vector/CometStructVector.java class CometStructVector (line 32) | public class CometStructVector extends CometDecodedVector { method CometStructVector (line 36) | public CometStructVector( method getChild (line 53) | @Override method slice (line 58) | @Override FILE: common/src/main/java/org/apache/comet/vector/CometVector.java class CometVector (line 45) | @IcebergApi method CometVector (line 64) | @IcebergApi method setNumNulls (line 74) | @IcebergApi method setNumValues (line 81) | @IcebergApi method numValues (line 85) | @IcebergApi method isFixedLength (line 89) | public boolean isFixedLength() { method getDecimal (line 93) | @Override method createDecimal (line 109) | private Decimal createDecimal(long unscaled, int precision, int scale) { method createDecimal (line 118) | private Decimal createDecimal(BigDecimal value, int precision, int sca... method getBinaryDecimal (line 130) | byte[] getBinaryDecimal(int i) { method copyBinaryDecimal (line 135) | public byte[] copyBinaryDecimal(int i, byte[] dest) { method getBoolean (line 152) | @Override method getByte (line 157) | @Override method getShort (line 162) | @Override method getInt (line 167) | @Override method getLong (line 172) | @Override method getLongDecimal (line 177) | public long getLongDecimal(int rowId) { method getFloat (line 181) | @Override method getDouble (line 186) | @Override method getUTF8String (line 191) | @Override method getBinary (line 196) | @Override method getArray (line 201) | @Override method getMap (line 206) | @Override method getChild (line 211) | @Override method close (line 216) | @Override method getDictionaryProvider (line 221) | public DictionaryProvider getDictionaryProvider() { method getValueVector (line 225) | @IcebergApi method slice (line 235) | @IcebergApi method getVector (line 245) | public static CometVector getVector( method getVector (line 271) | protected static CometVector getVector(ValueVector vector, boolean use... method notImplementedException (line 275) | private UnsupportedOperationException notImplementedException() { FILE: common/src/test/java/org/apache/comet/parquet/TestColumnReader.java class TestColumnReader (line 36) | public class TestColumnReader { method testIsFixedLength (line 37) | @Test FILE: common/src/test/java/org/apache/comet/parquet/TestCometInputFile.java class TestCometInputFile (line 25) | public class TestCometInputFile { method testIsAtLeastHadoop33 (line 26) | @Test FILE: common/src/test/java/org/apache/comet/parquet/TestFileReader.java class TestFileReader (line 79) | @SuppressWarnings("deprecation") method testEnableReadParallel (line 130) | @Test method testReadWrite (line 144) | @Test method testBloomFilterReadWrite (line 241) | @Test method testReadWriteDataPageV2 (line 287) | @Test method testColumnIndexFilter (line 395) | @Test method testColumnIndexReadWrite (line 488) | @Test method testWriteReadMergeScanRange (line 621) | @Test method addBloomFilter (line 634) | private void addBloomFilter(ParquetFileWriter w, String s, BloomFilter... method validateContains (line 642) | private void validateContains(PageReadStore pages, String[] path, int ... method validatePage (line 649) | private void validatePage(DataPage page, int values, BytesInput bytes)... method validateV2Page (line 654) | private void validateV2Page( method createStatistics (line 676) | private Statistics createStatistics(String min, String max, ColumnD... method assertStatsValuesEqual (line 684) | public static void assertStatsValuesEqual(Statistics expected, Stat... method statsC1 (line 696) | private Statistics statsC1(Binary... values) { method generateAndWriteData (line 714) | private HashMap generateAndWriteData( method readAndValidatePageData (line 754) | private void readAndValidatePageData( method testReadWrite (line 776) | public void testReadWrite(Configuration configuration, int numPages, i... class Generator (line 792) | static class Generator { method getString (line 799) | private byte[] getString(int minSize, int maxSize) { method generateValues (line 808) | private byte[] generateValues(int numValues, String type) throws IOE... method statsC2 (line 824) | private Statistics statsC2(Long... values) { FILE: common/src/test/java/org/apache/comet/parquet/TestUtils.java class TestUtils (line 34) | public class TestUtils { method testBuildColumnDescriptorWithTimestamp (line 36) | @Test method testBuildColumnDescriptorWithDecimal (line 70) | @Test method testBuildColumnDescriptorWithIntLogicalType (line 100) | @Test method testBuildColumnDescriptorWithStringLogicalType (line 128) | @Test FILE: dev/ci/check-suites.py function file_to_class_name (line 21) | def file_to_class_name(path: Path) -> str | None: FILE: dev/release/generate-changelog.py function print_pulls (line 25) | def print_pulls(repo_name, title, pulls): function generate_changelog (line 35) | def generate_changelog(repo, repo_name, tag1, tag2, version): function resolve_ref (line 145) | def resolve_ref(ref): function cli (line 157) | def cli(args=None): FILE: docs/generate-versions.py function get_major_minor_version (line 29) | def get_major_minor_version(version: str): function get_version_from_pom (line 33) | def get_version_from_pom(): function replace_in_files (line 46) | def replace_in_files(root: str, filename_pattern: str, search: str, repl... function insert_warning_after_asf_header (line 55) | def insert_warning_after_asf_header(root: str, warning: str): function get_user_guide_dir (line 69) | def get_user_guide_dir(major_minor: str): function publish_released_version (line 75) | def publish_released_version(version: str): function generate_docs (line 85) | def generate_docs(snapshot_version: str, latest_released_version: str, p... FILE: native/common/src/bin/analyze_trace.rs type TraceEvent (line 32) | struct TraceEvent { type MemorySnapshot (line 43) | struct MemorySnapshot { function format_bytes (line 49) | fn format_bytes(bytes: u64) -> String { function main (line 54) | fn main() { FILE: native/common/src/error.rs type SparkError (line 23) | pub enum SparkError { method to_json (line 199) | pub fn to_json(&self) -> String { method error_type_name (line 220) | pub(crate) fn error_type_name(&self) -> &'static str { method params_as_json (line 269) | pub(crate) fn params_as_json(&self) -> serde_json::Value { method exception_class (line 489) | pub fn exception_class(&self) -> &'static str { method error_class (line 554) | pub(crate) fn error_class(&self) -> Option<&'static str> { method from (line 730) | fn from(value: ArrowError) -> Self { type SparkResult (line 633) | pub type SparkResult = Result; function decimal_overflow_error (line 636) | pub fn decimal_overflow_error(value: i128, precision: u8, scale: i8) -> ... type SparkErrorWithContext (line 648) | pub struct SparkErrorWithContext { method new (line 657) | pub fn new(error: SparkError) -> Self { method with_context (line 665) | pub fn with_context(error: SparkError, context: Arc String { method fmt (line 706) | fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { method from (line 718) | fn from(error: SparkError) -> Self { method from (line 724) | fn from(value: SparkErrorWithContext) -> Self { method from (line 736) | fn from(value: SparkError) -> Self { function test_divide_by_zero_json (line 746) | fn test_divide_by_zero_json() { function test_remainder_by_zero_json (line 760) | fn test_remainder_by_zero_json() { function test_binary_overflow_json (line 769) | fn test_binary_overflow_json() { function test_invalid_array_index_json (line 789) | fn test_invalid_array_index_json() { function test_numeric_value_out_of_range_json (line 804) | fn test_numeric_value_out_of_range_json() { function test_cast_invalid_value_json (line 824) | fn test_cast_invalid_value_json() { function test_duplicated_map_key_json (line 841) | fn test_duplicated_map_key_json() { function test_null_map_key_json (line 854) | fn test_null_map_key_json() { function test_error_class_mapping (line 866) | fn test_error_class_mapping() { function test_exception_class_mapping (line 888) | fn test_exception_class_mapping() { FILE: native/common/src/query_context.rs type QueryContext (line 32) | pub struct QueryContext { method new (line 63) | pub fn new( method char_index_to_byte_offset (line 85) | fn char_index_to_byte_offset(&self, char_index: usize) -> Option { method format_summary (line 101) | pub fn format_summary(&self) -> String { method fragment (line 159) | fn fragment(&self) -> String { type QueryContextMap (line 187) | pub struct QueryContextMap { method new (line 193) | pub fn new() -> Self { method register (line 206) | pub fn register(&self, expr_id: u64, context: QueryContext) { method get (line 217) | pub fn get(&self, expr_id: u64) -> Option> { method clear (line 225) | pub fn clear(&self) { method len (line 231) | pub fn len(&self) -> usize { method is_empty (line 237) | pub fn is_empty(&self) -> bool { method default (line 243) | fn default() -> Self { function create_query_context_map (line 252) | pub fn create_query_context_map() -> Arc { function test_query_context_creation (line 261) | fn test_query_context_creation() { function test_query_context_serialization (line 282) | fn test_query_context_serialization() { function test_format_summary (line 300) | fn test_format_summary() { function test_format_summary_without_object (line 319) | fn test_format_summary_without_object() { function test_fragment (line 329) | fn test_fragment() { function test_arc_string_sharing (line 336) | fn test_arc_string_sharing() { function test_json_with_optional_fields (line 346) | fn test_json_with_optional_fields() { function test_map_register_and_get (line 357) | fn test_map_register_and_get() { function test_map_get_nonexistent (line 370) | fn test_map_get_nonexistent() { function test_map_clear (line 376) | fn test_map_clear() { function test_fragment_non_ascii_accented (line 393) | fn test_fragment_non_ascii_accented() { FILE: native/common/src/tracing.rs type Recorder (line 27) | pub struct Recorder { method new (line 39) | pub fn new() -> Self { method begin_task (line 58) | pub fn begin_task(&self, name: &str) { method end_task (line 62) | pub fn end_task(&self, name: &str) { method log_memory_usage (line 66) | pub fn log_memory_usage(&self, name: &str, usage_bytes: u64) { method log_event (line 78) | fn log_event(&self, name: &str, ph: &str) { method default (line 33) | fn default() -> Self { function get_thread_id (line 92) | pub fn get_thread_id() -> u64 { function trace_begin (line 101) | pub fn trace_begin(name: &str) { function trace_end (line 105) | pub fn trace_end(name: &str) { function log_memory_usage (line 109) | pub fn log_memory_usage(name: &str, value: u64) { function with_trace (line 113) | pub fn with_trace(label: &str, tracing_enabled: bool, f: F) -> T function with_trace_async (line 130) | pub async fn with_trace_async(label: &str, tracing_enabled: b... FILE: native/common/src/utils.rs function bytes_to_i128 (line 20) | pub fn bytes_to_i128(slice: &[u8]) -> i128 { FILE: native/core/benches/array_element_append.rs constant NUM_ELEMENTS (line 31) | const NUM_ELEMENTS: usize = 10000; function create_spark_unsafe_array_i32 (line 38) | fn create_spark_unsafe_array_i32(num_elements: usize, with_nulls: bool) ... function create_spark_unsafe_array_i64 (line 73) | fn create_spark_unsafe_array_i64(num_elements: usize, with_nulls: bool) ... function create_spark_unsafe_array_f64 (line 107) | fn create_spark_unsafe_array_f64(num_elements: usize, with_nulls: bool) ... function benchmark_array_conversion (line 140) | fn benchmark_array_conversion(c: &mut Criterion) { function config (line 263) | fn config() -> Criterion { FILE: native/core/benches/bit_util.rs function criterion_benchmark (line 34) | fn criterion_benchmark(c: &mut Criterion) { function bench_get_vlq_int (line 199) | fn bench_get_vlq_int(reader: &mut BitReader) { function config (line 205) | fn config() -> Criterion { FILE: native/core/benches/common.rs function seedable_rng (line 31) | pub fn seedable_rng() -> StdRng { function create_int64_array (line 35) | pub fn create_int64_array(size: usize, null_density: f32, min: i64, max:... function create_primitive_array (line 49) | pub fn create_primitive_array(size: usize, null_density: f32) -> Prim... function create_dictionary_array (line 69) | pub fn create_dictionary_array( FILE: native/core/benches/parquet_decode.rs function criterion_benchmark (line 22) | fn criterion_benchmark(c: &mut Criterion) { function config (line 47) | fn config() -> Criterion { FILE: native/core/benches/parquet_read.rs function bench (line 41) | fn bench(c: &mut Criterion) { function profiled (line 75) | fn profiled() -> Criterion { function build_test_schema (line 86) | fn build_test_schema() -> SchemaDescPtr { function seedable_rng (line 99) | fn seedable_rng() -> StdRng { constant NUM_PAGES (line 104) | const NUM_PAGES: usize = 1000; constant VALUES_PER_PAGE (line 105) | const VALUES_PER_PAGE: usize = 10_000; constant BATCH_SIZE (line 106) | const BATCH_SIZE: usize = 4096; function build_plain_int32_pages (line 108) | fn build_plain_int32_pages( type TestColumnReader (line 147) | struct TestColumnReader { method new (line 157) | pub fn new( method load_page (line 176) | fn load_page(&mut self) { type Item (line 186) | type Item = ArrayData; method next (line 188) | fn next(&mut self) -> Option { FILE: native/core/benches/perf.rs type FlamegraphProfiler (line 28) | pub struct FlamegraphProfiler<'a> { function new (line 34) | pub fn new(frequency: c_int) -> Self { method start_profiling (line 43) | fn start_profiling(&mut self, _benchmark_id: &str, _benchmark_dir: &Path) { method stop_profiling (line 47) | fn stop_profiling(&mut self, _benchmark_id: &str, benchmark_dir: &Path) { FILE: native/core/src/common/bit.rs function from_ne_slice (line 28) | pub fn from_ne_slice(bs: &[u8]) -> T { type FromBytes (line 38) | pub trait FromBytes: Sized { method from_le_bytes (line 40) | fn from_le_bytes(bs: Self::Buffer) -> Self; method from_be_bytes (line 41) | fn from_be_bytes(bs: Self::Buffer) -> Self; method from_ne_bytes (line 42) | fn from_ne_bytes(bs: Self::Buffer) -> Self; method from (line 43) | fn from(v: u64) -> Self; type Buffer (line 69) | type Buffer = [u8; 1]; method from_le_bytes (line 70) | fn from_le_bytes(bs: Self::Buffer) -> Self { method from_be_bytes (line 73) | fn from_be_bytes(bs: Self::Buffer) -> Self { method from_ne_bytes (line 76) | fn from_ne_bytes(bs: Self::Buffer) -> Self { method from (line 83) | fn from(v: u64) -> Self { function read_num_bytes_u64 (line 109) | pub fn read_num_bytes_u64(size: usize, src: &[u8]) -> u64 { function read_num_bytes_u32 (line 122) | pub fn read_num_bytes_u32(size: usize, src: &[u8]) -> u32 { function read_u64 (line 133) | pub fn read_u64(src: &[u8]) -> u64 { function read_u32 (line 139) | pub fn read_u32(src: &[u8]) -> u32 { function memcpy (line 145) | pub fn memcpy(source: &[u8], target: &mut [u8]) { function memcpy_value (line 153) | pub fn memcpy_value(source: &T, num_bytes: usize, target: &mut [u8]) function log2 (line 168) | pub fn log2(mut x: u64) -> u32 { function trailing_bits (line 178) | pub fn trailing_bits(v: u64, num_bits: usize) -> u64 { function set_bit (line 189) | pub fn set_bit(bits: &mut [u8], i: usize) { function set_bit_raw (line 198) | pub unsafe fn set_bit_raw(bits: *mut u8, i: usize) { function unset_bit (line 203) | pub fn unset_bit(bits: &mut [u8], i: usize) { function set_bits (line 208) | pub fn set_bits(bits: &mut [u8], offset: usize, length: usize) { function mix_hash (line 239) | pub fn mix_hash(lower: u64, upper: u64) -> u64 { function get_bit (line 248) | pub fn get_bit(data: &[u8], i: usize) -> bool { function get_bit_raw (line 257) | pub unsafe fn get_bit_raw(ptr: *const u8, i: usize) -> bool { type BitWriter (line 263) | pub struct BitWriter { method new (line 273) | pub fn new(max_bytes: usize) -> Self { method new_from_buf (line 286) | pub fn new_from_buf(buffer: Vec, start: usize) -> Self { method extend (line 301) | pub fn extend(&mut self, increment: usize) { method capacity (line 309) | pub fn capacity(&mut self) -> usize { method consume (line 315) | pub fn consume(mut self) -> Vec { method flush_buffer (line 324) | pub fn flush_buffer(&mut self) -> &[u8] { method clear (line 331) | pub fn clear(&mut self) { method flush (line 339) | pub fn flush(&mut self) { method skip (line 360) | pub fn skip(&mut self, num_bytes: usize) -> Result { method get_next_byte_ptr (line 380) | pub fn get_next_byte_ptr(&mut self, num_bytes: usize) -> Result<&mut [... method bytes_written (line 386) | pub fn bytes_written(&self) -> usize { method buffer (line 391) | pub fn buffer(&self) -> &[u8] { method byte_offset (line 396) | pub fn byte_offset(&self) -> usize { method buffer_len (line 404) | pub fn buffer_len(&self) -> usize { method write_at (line 409) | pub fn write_at(&mut self, offset: usize, value: u8) { method put_value (line 419) | pub fn put_value(&mut self, v: u64, num_bits: usize) -> bool { method put_aligned (line 454) | pub fn put_aligned(&mut self, val: T, num_bytes: usize) ->... method put_aligned_offset (line 474) | pub fn put_aligned_offset( method put_vlq_int (line 495) | pub fn put_vlq_int(&mut self, mut v: u64) -> bool { method put_zigzag_vlq_int (line 512) | pub fn put_zigzag_vlq_int(&mut self, v: i64) -> bool { constant MAX_VLQ_BYTE_LEN (line 520) | pub const MAX_VLQ_BYTE_LEN: usize = 10; type BitReader (line 522) | pub struct BitReader { method new (line 549) | pub fn new(buf: Buffer, len: usize) -> Self { method new_all (line 564) | pub fn new_all(buf: Buffer) -> Self { method reset (line 569) | pub fn reset(&mut self, buf: Buffer) { method get_byte_offset (line 583) | pub fn get_byte_offset(&self) -> usize { method get_value (line 590) | pub fn get_value(&mut self, num_bits: usize) -> Option { method get_u32_value (line 613) | pub fn get_u32_value(&mut self, num_bits: usize) -> u32 { method get_u64_value (line 618) | fn get_u64_value(&mut self, num_bits: usize) -> u64 { method get_bits (line 645) | pub fn get_bits(&mut self, dst: &mut [u8], offset: usize, num_bits: us... method get_bits_buffered (line 693) | fn get_bits_buffered(&mut self, dst: &mut [u8], offset: usize, num_bit... method skip_bits (line 728) | pub fn skip_bits(&mut self, num_bits: usize) -> usize { method get_u32_batch (line 779) | pub unsafe fn get_u32_batch(&mut self, mut dst: *mut u32, total: usize... method get_batch (line 808) | pub fn get_batch(&mut self, batch: &mut [T], num_bits: u... method get_aligned (line 889) | pub fn get_aligned(&mut self, num_bytes: usize) -> Optio... method get_vlq_int (line 932) | pub fn get_vlq_int(&mut self) -> Option { method get_zigzag_vlq_int (line 959) | pub fn get_zigzag_vlq_int(&mut self) -> Option { method reload_buffer_values (line 966) | fn reload_buffer_values(&mut self) { method from (line 978) | fn from(vec: Vec) -> Self { function round_upto_power_of_2 (line 988) | pub fn round_upto_power_of_2(num: usize, factor: usize) -> usize { function test_read_num_bytes_u64 (line 1005) | fn test_read_num_bytes_u64() { function test_read_u64 (line 1016) | fn test_read_u64() { function test_read_num_bytes_u32 (line 1022) | fn test_read_num_bytes_u32() { function test_read_u32 (line 1033) | fn test_read_u32() { function test_bit_reader_get_byte_offset (line 1039) | fn test_bit_reader_get_byte_offset() { function test_bit_reader_get_value (line 1054) | fn test_bit_reader_get_value() { function test_bit_reader_get_value_boundary (line 1064) | fn test_bit_reader_get_value_boundary() { function test_bit_reader_get_aligned (line 1074) | fn test_bit_reader_get_aligned() { function test_bit_reader_get_vlq_int (line 1086) | fn test_bit_reader_get_vlq_int() { function test_bit_reader_get_zigzag_vlq_int (line 1095) | fn test_bit_reader_get_zigzag_vlq_int() { function test_set_bit (line 1105) | fn test_set_bit() { function test_set_bits (line 1124) | fn test_set_bits() { function test_get_bit (line 1139) | fn test_get_bit() { function test_log2 (line 1166) | fn test_log2() { function test_skip (line 1180) | fn test_skip() { function test_get_next_byte_ptr (line 1194) | fn test_get_next_byte_ptr() { function test_consume_flush_buffer (line 1208) | fn test_consume_flush_buffer() { function test_put_get_bool (line 1221) | fn test_put_get_bool() { function test_put_value_roundtrip (line 1272) | fn test_put_value_roundtrip() { function test_put_value_rand_numbers (line 1285) | fn test_put_value_rand_numbers(total: usize, num_bits: usize) { function test_get_bits (line 1314) | fn test_get_bits() { function test_skip_bits (line 1364) | fn test_skip_bits() { function test_get_batch (line 1407) | fn test_get_batch() { function test_get_batch_helper (line 1420) | fn test_get_batch_helper(total: usize, num_bits: usize) function test_get_u32_batch (line 1454) | fn test_get_u32_batch() { function test_put_aligned_roundtrip (line 1484) | fn test_put_aligned_roundtrip() { function test_put_aligned_rand_numbers (line 1495) | fn test_put_aligned_rand_numbers(total: usize, num_bits: usize) function test_put_vlq_int (line 1554) | fn test_put_vlq_int() { function test_put_zigzag_vlq_int (line 1579) | fn test_put_zigzag_vlq_int() { FILE: native/core/src/common/buffer.rs type CometBuffer (line 41) | pub struct CometBuffer { method new (line 59) | pub fn new(capacity: usize) -> Self { method capacity (line 75) | pub fn capacity(&self) -> usize { method len (line 80) | pub fn len(&self) -> usize { method is_empty (line 85) | pub fn is_empty(&self) -> bool { method as_slice (line 90) | pub fn as_slice(&self) -> &[u8] { method as_slice_mut (line 95) | pub fn as_slice_mut(&mut self) -> &mut [u8] { method as_ptr (line 103) | pub const fn as_ptr(&self) -> *const u8 { method as_mut_ptr (line 110) | pub fn as_mut_ptr(&mut self) -> *mut u8 { method to_arrow (line 132) | pub unsafe fn to_arrow(&self) -> Result { method check_reference (line 146) | pub fn check_reference(&self) -> Result<(), ExecutionError> { method reset (line 157) | pub fn reset(&mut self) { method resize (line 167) | pub fn resize(&mut self, new_capacity: usize) { method reallocate (line 181) | unsafe fn reallocate( type Target (line 229) | type Target = [u8]; method deref (line 231) | fn deref(&self) -> &[u8] { method deref_mut (line 237) | fn deref_mut(&mut self) -> &mut [u8] { method from_ptr (line 258) | pub fn from_ptr(ptr: *const u8, len: usize, capacity: usize) -> Self { method extend_from_slice (line 279) | pub fn extend_from_slice(&mut self, offset: usize, src: &[u8]) { constant ALIGNMENT (line 55) | const ALIGNMENT: usize = 64; method drop (line 201) | fn drop(&mut self) { method eq (line 214) | fn eq(&self, other: &CometBuffer) -> bool { type CometBufferAllocation (line 244) | struct CometBufferAllocation {} method new (line 247) | fn new() -> Self { function test_buffer_new (line 297) | fn test_buffer_new() { function test_resize (line 305) | fn test_resize() { function test_extend_from_slice (line 321) | fn test_extend_from_slice() { function test_to_arrow (line 335) | fn test_to_arrow() { function test_unowned (line 353) | fn test_unowned() { FILE: native/core/src/execution/columnar_to_row.rs constant MAX_LONG_DIGITS (line 49) | const MAX_LONG_DIGITS: u8 = 18; function write_bytes_padded (line 142) | fn write_bytes_padded(buffer: &mut Vec, bytes: &[u8]) -> usize { type TypedArray (line 153) | enum TypedArray<'a> { function from_array (line 183) | fn from_array(array: &'a ArrayRef) -> CometResult { function is_null (line 255) | fn is_null(&self, row_idx: usize) -> bool { function get_fixed_value (line 287) | fn get_fixed_value(&self, row_idx: usize) -> i64 { function is_variable_length (line 309) | fn is_variable_length(&self) -> bool { function write_variable_to_buffer (line 327) | fn write_variable_to_buffer(&self, buffer: &mut Vec, row_idx: usize)... type TypedElements (line 373) | enum TypedElements<'a> { function from_array (line 395) | fn from_array(array: &'a ArrayRef, element_type: &DataType) -> Self { function element_size (line 437) | fn element_size(&self) -> usize { function supports_bulk_copy (line 452) | fn supports_bulk_copy(&self) -> bool { function is_null_at (line 468) | fn is_null_at(&self, idx: usize) -> bool { function is_fixed_width (line 495) | fn is_fixed_width(&self) -> bool { function get_fixed_value (line 513) | fn get_fixed_value(&self, idx: usize) -> i64 { function write_variable_value (line 530) | fn write_variable_value(&self, buffer: &mut Vec, idx: usize) -> Come... function write_range_to_buffer (line 552) | fn write_range_to_buffer( function bulk_copy_range (line 599) | fn bulk_copy_range( function write_elements_slow (line 664) | fn write_elements_slow( function set_null_bit (line 808) | fn set_null_bit(buffer: &mut [u8], null_bitset_start: usize, idx: usize) { function is_fixed_width (line 820) | fn is_fixed_width(data_type: &DataType) -> bool { function is_all_fixed_width (line 838) | fn is_all_fixed_width(schema: &[DataType]) -> bool { type ColumnarToRowContext (line 847) | pub struct ColumnarToRowContext { method new (line 874) | pub fn new(schema: Vec, batch_size: usize) -> Self { method calculate_bitset_width (line 904) | pub const fn calculate_bitset_width(num_fields: usize) -> usize { method round_up_to_8 (line 910) | const fn round_up_to_8(value: usize) -> usize { method convert (line 927) | pub fn convert( method maybe_cast_to_schema_type (line 998) | fn maybe_cast_to_schema_type( method convert_fixed_width (line 1084) | fn convert_fixed_width( method write_column_fixed_width (line 1123) | fn write_column_fixed_width( method write_row_typed (line 1258) | fn write_row_typed( method buffer_ptr (line 1322) | pub fn buffer_ptr(&self) -> *const u8 { method schema (line 1327) | pub fn schema(&self) -> &[DataType] { function get_field_value (line 1334) | fn get_field_value(data_type: &DataType, array: &ArrayRef, row_idx: usiz... function write_dictionary_to_buffer (line 1414) | fn write_dictionary_to_buffer( function write_dictionary_to_buffer_with_key (line 1455) | fn write_dictionary_to_buffer_with_key( function i128_to_spark_decimal_bytes (line 1507) | fn i128_to_spark_decimal_bytes(value: i128) -> Vec { function round_up_to_8 (line 1536) | const fn round_up_to_8(value: usize) -> usize { function write_array_element (line 1542) | fn write_array_element(buffer: &mut [u8], data_type: &DataType, value: i... function write_struct_to_buffer_typed (line 1577) | fn write_struct_to_buffer_typed( function write_struct_to_buffer (line 1628) | fn write_struct_to_buffer( function write_list_to_buffer (line 1712) | fn write_list_to_buffer( function write_large_list_to_buffer (line 1738) | fn write_large_list_to_buffer( function write_map_to_buffer (line 1763) | fn write_map_to_buffer( function write_nested_variable_to_buffer (line 1816) | fn write_nested_variable_to_buffer( function test_bitset_width_calculation (line 1895) | fn test_bitset_width_calculation() { function test_round_up_to_8 (line 1905) | fn test_round_up_to_8() { function test_convert_int_array (line 1914) | fn test_convert_int_array() { function test_convert_multiple_columns (line 1934) | fn test_convert_multiple_columns() { function test_fixed_width_fast_path (line 1956) | fn test_fixed_width_fast_path() { function test_mixed_schema_uses_general_path (line 2015) | fn test_mixed_schema_uses_general_path() { function test_convert_string_array (line 2028) | fn test_convert_string_array() { function test_i128_to_spark_decimal_bytes (line 2048) | fn test_i128_to_spark_decimal_bytes() { function test_list_data_conversion (line 2063) | fn test_list_data_conversion() { function test_list_data_conversion_multiple_rows (line 2102) | fn test_list_data_conversion_multiple_rows() { function test_map_data_conversion (line 2148) | fn test_map_data_conversion() { function test_map_data_conversion_multiple_rows (line 2235) | fn test_map_data_conversion_multiple_rows() { function test_map_data_conversion_sliced_maparray (line 2343) | fn test_map_data_conversion_sliced_maparray() { function test_large_list_data_conversion (line 2423) | fn test_large_list_data_conversion() { function test_convert_fixed_size_binary_array (line 2464) | fn test_convert_fixed_size_binary_array() { function test_convert_dictionary_decimal_array (line 2503) | fn test_convert_dictionary_decimal_array() { function test_convert_int32_to_decimal128 (line 2549) | fn test_convert_int32_to_decimal128() { function test_convert_int64_to_decimal128 (line 2587) | fn test_convert_int64_to_decimal128() { FILE: native/core/src/execution/expressions/arithmetic.rs type CheckedBinaryExpr (line 34) | pub struct CheckedBinaryExpr { method new (line 42) | pub fn new(child: Arc, query_context: Option bool { method fmt (line 51) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method eq (line 57) | fn eq(&self, other: &Self) -> bool { method hash (line 74) | fn hash(&self, state: &mut H) { method as_any (line 80) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 84) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 88) | fn data_type(&self, input_schema: &Schema) -> datafusion::common::Result... method nullable (line 92) | fn nullable(&self, input_schema: &Schema) -> datafusion::common::Result<... method evaluate (line 96) | fn evaluate(&self, batch: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method with_new_children (line 120) | fn with_new_children( type IntegralDivideBuilder (line 202) | pub struct IntegralDivideBuilder; method build (line 205) | fn build( type RemainderBuilder (line 229) | pub struct RemainderBuilder; method build (line 232) | fn build( type UnaryMinusBuilder (line 259) | pub struct UnaryMinusBuilder; method build (line 262) | fn build( FILE: native/core/src/execution/expressions/partition.rs type SparkPartitionIdBuilder (line 29) | pub struct SparkPartitionIdBuilder; method build (line 32) | fn build( type MonotonicallyIncreasingIdBuilder (line 44) | pub struct MonotonicallyIncreasingIdBuilder; method build (line 47) | fn build( FILE: native/core/src/execution/expressions/random.rs type RandBuilder (line 28) | pub struct RandBuilder; method build (line 31) | fn build( type RandnBuilder (line 43) | pub struct RandnBuilder; method build (line 46) | fn build( FILE: native/core/src/execution/expressions/strings.rs type SubstringBuilder (line 38) | pub struct SubstringBuilder; method build (line 41) | fn build( type LikeBuilder (line 63) | pub struct LikeBuilder; method build (line 66) | fn build( type RlikeBuilder (line 81) | pub struct RlikeBuilder; method build (line 84) | fn build( type FromJsonBuilder (line 103) | pub struct FromJsonBuilder; method build (line 106) | fn build( FILE: native/core/src/execution/expressions/subquery.rs type Subquery (line 39) | pub struct Subquery { method new (line 50) | pub fn new(exec_context_id: i64, id: i64, data_type: DataType) -> Self { method fmt (line 60) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 66) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 70) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 74) | fn data_type(&self, _: &Schema) -> datafusion::common::Result { method nullable (line 78) | fn nullable(&self, _: &Schema) -> datafusion::common::Result { method evaluate (line 82) | fn evaluate(&self, _: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method with_new_children (line 196) | fn with_new_children( FILE: native/core/src/execution/expressions/temporal.rs type HourBuilder (line 38) | pub struct HourBuilder; method build (line 41) | fn build( type MinuteBuilder (line 65) | pub struct MinuteBuilder; method build (line 68) | fn build( type SecondBuilder (line 92) | pub struct SecondBuilder; method build (line 95) | fn build( type UnixTimestampBuilder (line 119) | pub struct UnixTimestampBuilder; method build (line 122) | fn build( type TruncTimestampBuilder (line 147) | pub struct TruncTimestampBuilder; method build (line 150) | fn build( type HoursTransformBuilder (line 165) | pub struct HoursTransformBuilder; method build (line 168) | fn build( FILE: native/core/src/execution/jni_api.rs function log_jemalloc_usage (line 112) | fn log_jemalloc_usage() { type ThreadPoolMap (line 121) | type ThreadPoolMap = HashMap>>; function get_thread_memory_pools (line 125) | fn get_thread_memory_pools() -> &'static Mutex { function register_memory_pool (line 129) | fn register_memory_pool(thread_id: u64, context_id: i64, pool: Arc usize { function total_reserved_for_thread (line 158) | fn total_reserved_for_thread(thread_id: u64) -> usize { function parse_usize_env_var (line 176) | fn parse_usize_env_var(name: &str) -> Option { function build_runtime (line 180) | fn build_runtime(default_worker_threads: Option) -> Runtime { function init_runtime (line 202) | pub fn init_runtime(default_worker_threads: usize) { function get_runtime (line 207) | pub fn get_runtime() -> &'static Runtime { function op_name (line 212) | fn op_name(op: &OpStruct) -> &'static str { function build_tracing_event_name (line 235) | fn build_tracing_event_name(plan: &Operator) -> String { function collect_op_names (line 248) | fn collect_op_names<'a>(op: &'a Operator, names: &mut std::collections::... type ExecutionContext (line 258) | struct ExecutionContext { function Java_org_apache_comet_Native_createPlan (line 311) | pub unsafe extern "system" fn Java_org_apache_comet_Native_createPlan( function prepare_datafusion_session_context (line 486) | fn prepare_datafusion_session_context( function register_datafusion_spark_function (line 542) | fn register_datafusion_spark_function(session_ctx: &SessionContext) { function prepare_output (line 571) | fn prepare_output( function pull_input_batches (line 648) | fn pull_input_batches(exec_context: &mut ExecutionContext) -> Result<(),... function Java_org_apache_comet_Native_executePlan (line 664) | pub unsafe extern "system" fn Java_org_apache_comet_Native_executePlan( function Java_org_apache_comet_Native_releasePlan (line 855) | pub extern "system" fn Java_org_apache_comet_Native_releasePlan( function update_metrics (line 886) | fn update_metrics(env: &mut Env, exec_context: &mut ExecutionContext) ->... function log_plan_metrics (line 895) | fn log_plan_metrics(exec_context: &ExecutionContext, stage_id: jint, par... function convert_datatype_arrays (line 910) | fn convert_datatype_arrays( function get_execution_context (line 929) | fn get_execution_context<'a>(id: i64) -> &'a mut ExecutionContext { function Java_org_apache_comet_Native_writeSortedFileNative (line 941) | pub unsafe extern "system" fn Java_org_apache_comet_Native_writeSortedFi... function Java_org_apache_comet_Native_sortRowPartitionsNative (line 1023) | pub extern "system" fn Java_org_apache_comet_Native_sortRowPartitionsNat... function Java_org_apache_comet_Native_decodeShuffleBlock (line 1056) | pub unsafe extern "system" fn Java_org_apache_comet_Native_decodeShuffle... function Java_org_apache_comet_Native_traceBegin (line 1079) | pub unsafe extern "system" fn Java_org_apache_comet_Native_traceBegin( function Java_org_apache_comet_Native_traceEnd (line 1094) | pub unsafe extern "system" fn Java_org_apache_comet_Native_traceEnd( function Java_org_apache_comet_Native_logMemoryUsage (line 1109) | pub unsafe extern "system" fn Java_org_apache_comet_Native_logMemoryUsage( function Java_org_apache_comet_Native_getRustThreadId (line 1125) | pub extern "system" fn Java_org_apache_comet_Native_getRustThreadId( function Java_org_apache_comet_Native_columnarToRowInit (line 1145) | pub unsafe extern "system" fn Java_org_apache_comet_Native_columnarToRow... function Java_org_apache_comet_Native_columnarToRowConvert (line 1167) | pub unsafe extern "system" fn Java_org_apache_comet_Native_columnarToRow... function Java_org_apache_comet_Native_columnarToRowClose (line 1255) | pub unsafe extern "system" fn Java_org_apache_comet_Native_columnarToRow... FILE: native/core/src/execution/memory_pools/config.rs type MemoryPoolType (line 21) | pub(crate) enum MemoryPoolType { method is_task_shared (line 34) | pub(crate) fn is_task_shared(&self) -> bool { type MemoryPoolConfig (line 45) | pub(crate) struct MemoryPoolConfig { method new (line 51) | pub(crate) fn new(pool_type: MemoryPoolType, pool_size: usize) -> Self { function parse_memory_pool_config (line 59) | pub(crate) fn parse_memory_pool_config( FILE: native/core/src/execution/memory_pools/fair_pool.rs type CometFairMemoryPool (line 36) | pub struct CometFairMemoryPool { method new (line 59) | pub fn new( method acquire (line 70) | fn acquire(&self, additional: usize) -> CometResult { method release (line 78) | fn release(&self, size: usize) -> CometResult<()> { type CometFairPoolState (line 42) | struct CometFairPoolState { method fmt (line 48) | fn fmt(&self, f: &mut Formatter<'_>) -> FmtResult { method register (line 90) | fn register(&self, _: &MemoryConsumer) { method unregister (line 98) | fn unregister(&self, _: &MemoryConsumer) { method grow (line 106) | fn grow(&self, _reservation: &MemoryReservation, additional: usize) { method shrink (line 110) | fn shrink(&self, _reservation: &MemoryReservation, subtractive: usize) { method try_grow (line 128) | fn try_grow( method reserved (line 172) | fn reserved(&self) -> usize { FILE: native/core/src/execution/memory_pools/logging_pool.rs type LoggingMemoryPool (line 25) | pub(crate) struct LoggingMemoryPool { method new (line 31) | pub fn new(task_attempt_id: u64, pool: Arc) -> Self { method register (line 40) | fn register(&self, consumer: &MemoryConsumer) { method unregister (line 49) | fn unregister(&self, consumer: &MemoryConsumer) { method grow (line 58) | fn grow(&self, reservation: &MemoryReservation, additional: usize) { method shrink (line 68) | fn shrink(&self, reservation: &MemoryReservation, shrink: usize) { method try_grow (line 78) | fn try_grow( method reserved (line 105) | fn reserved(&self) -> usize { method memory_limit (line 109) | fn memory_limit(&self) -> MemoryLimit { FILE: native/core/src/execution/memory_pools/mod.rs function create_memory_pool (line 37) | pub(crate) fn create_memory_pool( FILE: native/core/src/execution/memory_pools/task_shared.rs type PerTaskMemoryPool (line 28) | pub(crate) struct PerTaskMemoryPool { method new (line 34) | pub(crate) fn new(memory_pool: Arc) -> Self { function handle_task_shared_pool_release (line 44) | pub(crate) fn handle_task_shared_pool_release(pool_type: MemoryPoolType,... FILE: native/core/src/execution/memory_pools/unified_pool.rs type CometUnifiedMemoryPool (line 37) | pub struct CometUnifiedMemoryPool { method new (line 52) | pub fn new( method acquire_from_spark (line 64) | fn acquire_from_spark(&self, additional: usize) -> CometResult { method release_to_spark (line 73) | fn release_to_spark(&self, size: usize) -> CometResult<()> { method fmt (line 44) | fn fmt(&self, f: &mut Formatter<'_>) -> FmtResult { method drop (line 82) | fn drop(&mut self) { method grow (line 97) | fn grow(&self, reservation: &MemoryReservation, additional: usize) { method shrink (line 101) | fn shrink(&self, _: &MemoryReservation, size: usize) { method try_grow (line 119) | fn try_grow(&self, _: &MemoryReservation, additional: usize) -> Result<(... method reserved (line 151) | fn reserved(&self) -> usize { FILE: native/core/src/execution/metrics/utils.rs function update_comet_metric (line 30) | pub(crate) fn update_comet_metric( function to_native_metric_node (line 45) | pub(crate) fn to_native_metric_node( FILE: native/core/src/execution/mod.rs function it_works (line 38) | fn it_works() { FILE: native/core/src/execution/operators/copy.rs type CopyMode (line 26) | pub enum CopyMode { function copy_array (line 34) | pub(crate) fn copy_array(array: &dyn Array) -> ArrayRef { function copy_or_unpack_array (line 70) | pub(crate) fn copy_or_unpack_array( FILE: native/core/src/execution/operators/csv_scan.rs function init_csv_datasource_exec (line 32) | pub fn init_csv_datasource_exec( function build_csv_source (line 55) | fn build_csv_source(schema: SchemaRef, options: &CsvOptions) -> Result Result { FILE: native/core/src/execution/operators/expand.rs type ExpandExec (line 41) | pub struct ExpandExec { method new (line 50) | pub fn new( method fmt_as (line 72) | fn fmt_as(&self, t: DisplayFormatType, f: &mut std::fmt::Formatter) -> s... method as_any (line 94) | fn as_any(&self) -> &dyn Any { method schema (line 98) | fn schema(&self) -> SchemaRef { method children (line 102) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 106) | fn with_new_children( method execute (line 118) | fn execute( method properties (line 132) | fn properties(&self) -> &Arc { method name (line 136) | fn name(&self) -> &str { type ExpandStream (line 141) | pub struct ExpandStream { method new (line 152) | pub fn new( method expand (line 168) | fn expand( type Item (line 189) | type Item = datafusion::common::Result; method poll_next (line 191) | fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll SchemaRef { FILE: native/core/src/execution/operators/iceberg_scan.rs type IcebergScanExec (line 55) | pub struct IcebergScanExec { method new (line 73) | pub fn new( method compute_properties (line 96) | fn compute_properties(schema: SchemaRef, num_partitions: usize) -> Arc... method execute_with_tasks (line 150) | fn execute_with_tasks( method storage_factory_for (line 195) | fn storage_factory_for(path: &str) -> Result, ... method load_file_io (line 215) | fn load_file_io( method name (line 107) | fn name(&self) -> &str { method as_any (line 111) | fn as_any(&self) -> &dyn Any { method schema (line 115) | fn schema(&self) -> SchemaRef { method properties (line 119) | fn properties(&self) -> &Arc { method children (line 123) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 127) | fn with_new_children( method execute (line 134) | fn execute( method metrics (line 142) | fn metrics(&self) -> Option { type IcebergScanMetrics (line 231) | struct IcebergScanMetrics { method new (line 239) | fn new(metrics: &ExecutionPlanMetricsSet) -> Self { type IcebergStreamWrapper (line 250) | struct IcebergStreamWrapper { type CachedProjection (line 263) | struct CachedProjection { type Item (line 272) | type Item = DFResult; method poll_next (line 274) | fn poll_next(mut self: Pin<&mut Self>, cx: &mut Context<'_>) -> Poll SchemaRef { method fmt_as (line 331) | fn fmt_as(&self, _t: DisplayFormatType, f: &mut fmt::Formatter) -> fmt::... function build_projection_expressions (line 345) | fn build_projection_expressions( function adapt_batch_with_expressions (line 367) | fn adapt_batch_with_expressions( FILE: native/core/src/execution/operators/parquet_writer.rs type ParquetWriter (line 62) | enum ParquetWriter { method write (line 80) | async fn write( method close (line 135) | async fn close(self) -> std::result::Result<(), parquet::errors::Parqu... type ParquetWriterExec (line 194) | pub struct ParquetWriterExec { method try_new (line 222) | pub fn try_new( method compression_to_parquet (line 258) | fn compression_to_parquet(&self) -> Result { method create_arrow_writer (line 279) | fn create_arrow_writer( method fmt_as (line 391) | fn fmt_as(&self, t: DisplayFormatType, f: &mut Formatter) -> fmt::Result { method as_any (line 407) | fn as_any(&self) -> &dyn Any { method name (line 411) | fn name(&self) -> &str { method metrics (line 415) | fn metrics(&self) -> Option { method properties (line 419) | fn properties(&self) -> &Arc { method schema (line 423) | fn schema(&self) -> SchemaRef { method children (line 427) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 431) | fn with_new_children( method execute (line 453) | fn execute( function create_test_record_batch (line 587) | fn create_test_record_batch(batch_id: i32) -> Result { function test_write_to_hdfs_sync (line 612) | async fn test_write_to_hdfs_sync() -> Result<()> { function test_write_to_hdfs_streaming (line 663) | async fn test_write_to_hdfs_streaming() -> Result<()> { function test_parquet_writer_streaming (line 744) | async fn test_parquet_writer_streaming() -> Result<()> { function test_parquet_writer_exec_with_memory_input (line 797) | async fn test_parquet_writer_exec_with_memory_input() -> Result<()> { FILE: native/core/src/execution/operators/projection.rs type ProjectionBuilder (line 35) | pub struct ProjectionBuilder; method build (line 38) | fn build( FILE: native/core/src/execution/operators/scan.rs type ScanExec (line 57) | pub struct ScanExec { method new (line 83) | pub fn new( method unpack_dictionary_type (line 121) | fn unpack_dictionary_type(dt: &DataType) -> DataType { method set_input_batch (line 130) | pub fn set_input_batch(&mut self, input: InputBatch) { method get_next_batch (line 135) | pub fn get_next_batch(&mut self) -> Result<(), CometError> { method get_next (line 159) | fn get_next( method allocate_and_fetch_batch (line 262) | fn allocate_and_fetch_batch( method get_selection_indices (line 309) | fn get_selection_indices( function schema_from_data_types (line 371) | fn schema_from_data_types(data_types: &[DataType]) -> SchemaRef { method as_any (line 386) | fn as_any(&self) -> &dyn Any { method schema (line 390) | fn schema(&self) -> SchemaRef { method children (line 394) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 398) | fn with_new_children( method execute (line 405) | fn execute( method properties (line 418) | fn properties(&self) -> &Arc { method name (line 422) | fn name(&self) -> &str { method metrics (line 426) | fn metrics(&self) -> Option { method fmt_as (line 432) | fn fmt_as(&self, t: DisplayFormatType, f: &mut std::fmt::Formatter) -> s... type ScanStream (line 451) | struct ScanStream<'a> { function new (line 465) | pub fn new( function build_record_batch (line 481) | fn build_record_batch( type Item (line 514) | type Item = DataFusionResult; method poll_next (line 516) | fn poll_next(self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll SchemaRef { type InputBatch (line 553) | pub enum InputBatch { method new (line 567) | pub fn new(columns: Vec, num_rows: Option) -> Self { FILE: native/core/src/execution/operators/shuffle_scan.rs type ShuffleScanExec (line 52) | pub struct ShuffleScanExec { method new (line 74) | pub fn new( method set_input_batch (line 106) | pub fn set_input_batch(&mut self, input: InputBatch) { method get_next_batch (line 112) | pub fn get_next_batch(&mut self) -> Result<(), CometError> { method get_next (line 136) | fn get_next( function unpack_dictionary (line 205) | fn unpack_dictionary(array: &ArrayRef) -> ArrayRef { function schema_from_data_types (line 213) | fn schema_from_data_types(data_types: &[DataType]) -> SchemaRef { method as_any (line 224) | fn as_any(&self) -> &dyn Any { method schema (line 228) | fn schema(&self) -> SchemaRef { method children (line 232) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 236) | fn with_new_children( method execute (line 243) | fn execute( method properties (line 255) | fn properties(&self) -> &Arc { method name (line 259) | fn name(&self) -> &str { method metrics (line 263) | fn metrics(&self) -> Option { method fmt_as (line 269) | fn fmt_as(&self, t: DisplayFormatType, f: &mut std::fmt::Formatter) -> s... type ShuffleScanStream (line 287) | struct ShuffleScanStream { method new (line 295) | pub fn new( type Item (line 308) | type Item = DataFusionResult; method poll_next (line 310) | fn poll_next(self: Pin<&mut Self>, _: &mut Context<'_>) -> Poll SchemaRef { function test_read_compressed_ipc_block (line 366) | fn test_read_compressed_ipc_block() { function test_dictionary_encoded_shuffle_block_is_unpacked (line 412) | fn test_dictionary_encoded_shuffle_block_is_unpacked() { FILE: native/core/src/execution/planner.rs type PhyAggResult (line 138) | type PhyAggResult = Result, ExecutionError>; type PhyExprResult (line 139) | type PhyExprResult = Result, String)>, Execut... type PartitionPhyExprResult (line 140) | type PartitionPhyExprResult = Result>, Executi... type PlanCreationResult (line 141) | pub type PlanCreationResult = type JoinParameters (line 144) | struct JoinParameters { type BinaryExprOptions (line 153) | pub struct BinaryExprOptions { constant TEST_EXEC_CONTEXT_ID (line 157) | pub const TEST_EXEC_CONTEXT_ID: i64 = -1; type PhysicalPlanner (line 160) | pub struct PhysicalPlanner { method new (line 175) | pub fn new(session_ctx: Arc, partition: i32) -> Self { method with_exec_id (line 184) | pub fn with_exec_id(self, exec_context_id: i64) -> Self { method session_ctx (line 194) | pub fn session_ctx(&self) -> &Arc { method partition (line 199) | pub fn partition(&self) -> i32 { method get_partitioned_files (line 204) | fn get_partitioned_files( method create_expr (line 256) | pub(crate) fn create_expr( method create_sort_expr (line 690) | fn create_sort_expr<'a>( method create_binary_expr (line 716) | pub fn create_binary_expr( method create_binary_expr_with_options (line 739) | pub fn create_binary_expr_with_options( method create_plan (line 915) | pub(crate) fn create_plan<'a>( method parse_join_parameters (line 1803) | fn parse_join_parameters( method create_agg_expr (line 1946) | fn create_agg_expr( method create_window_expr (line 2273) | fn create_window_expr<'a>( method process_agg_func (line 2447) | fn process_agg_func( method find_df_window_function (line 2488) | fn find_df_window_function(&self, name: &str) -> Option Self { function expr_to_columns (line 2765) | fn expr_to_columns( type JoinFilterRewriter (line 2799) | struct JoinFilterRewriter<'a> { function new (line 2807) | fn new<'a>( type Node (line 2823) | type Node = Arc; method f_down (line 2825) | fn f_down(&mut self, node: Self::Node) -> datafusion::common::Result Result Option Option bool { function test_unpack_dictionary_primitive (line 3764) | fn test_unpack_dictionary_primitive() { constant STRING_TYPE_ID (line 3827) | const STRING_TYPE_ID: i32 = 7; function test_unpack_dictionary_string (line 3830) | fn test_unpack_dictionary_string() { function to_datafusion_filter (line 3906) | async fn to_datafusion_filter() { function from_datafusion_error_to_comet (line 3929) | async fn from_datafusion_error_to_comet() { function create_filter (line 3938) | fn create_filter(child_op: spark_operator::Operator, value: i32) -> spar... function create_filter_literal (line 3951) | fn create_filter_literal( function spark_plan_metrics_filter (line 3992) | fn spark_plan_metrics_filter() { function spark_plan_metrics_hash_join (line 4006) | fn spark_plan_metrics_hash_join() { function create_bound_reference (line 4031) | fn create_bound_reference(index: i32) -> Expr { function create_scan (line 4042) | fn create_scan() -> Operator { function create_proto_datatype (line 4054) | fn create_proto_datatype() -> spark_expression::DataType { function test_create_array (line 4062) | fn test_create_array() { function test_array_repeat (line 4189) | fn test_array_repeat() { function make_parquet_data (line 4326) | async fn make_parquet_data( function test_nested_types_list_of_struct_by_index (line 4390) | async fn test_nested_types_list_of_struct_by_index() -> Result<(), DataF... function test_nested_types_map_keys (line 4431) | async fn test_nested_types_map_keys() -> Result<(), DataFusionError> { function test_nested_types_extract_missing_struct_names_non_overlap (line 4482) | async fn test_nested_types_extract_missing_struct_names_non_overlap( function test_nested_types_extract_missing_struct_names_single_field (line 4501) | async fn test_nested_types_extract_missing_struct_names_single_field( function test_nested_types_extract_missing_struct_names_missing_field (line 4523) | async fn test_nested_types_extract_missing_struct_names_missing_field( function test_literal_to_list (line 4547) | async fn test_literal_to_list() -> Result<(), DataFusionError> { function test_date_sub_with_int8_cast_error (line 4676) | fn test_date_sub_with_int8_cast_error() { FILE: native/core/src/execution/planner/expression_registry.rs type ExpressionBuilder (line 30) | pub trait ExpressionBuilder: Send + Sync { method build (line 32) | fn build( type ExpressionType (line 42) | pub enum ExpressionType { type ExpressionRegistry (line 117) | pub struct ExpressionRegistry { method new (line 123) | fn new() -> Self { method global (line 133) | pub fn global() -> &'static ExpressionRegistry { method can_handle (line 139) | pub fn can_handle(&self, spark_expr: &Expr) -> bool { method create_expr (line 148) | pub fn create_expr( method register_all_expressions (line 167) | fn register_all_expressions(&mut self) { method register_arithmetic_expressions (line 197) | fn register_arithmetic_expressions(&mut self) { method register_comparison_expressions (line 219) | fn register_comparison_expressions(&mut self) { method register_bitwise_expressions (line 241) | fn register_bitwise_expressions(&mut self) { method register_logical_expressions (line 261) | fn register_logical_expressions(&mut self) { method register_null_check_expressions (line 273) | fn register_null_check_expressions(&mut self) { method register_string_expressions (line 283) | fn register_string_expressions(&mut self) { method register_temporal_expressions (line 297) | fn register_temporal_expressions(&mut self) { method get_expression_type (line 321) | fn get_expression_type(spark_expr: &Expr) -> Result Self { method global (line 77) | pub fn global() -> &'static OperatorRegistry { method can_handle (line 87) | pub fn can_handle(&self, spark_operator: &Operator) -> bool { method create_plan (line 94) | pub fn create_plan( method register_all_operators (line 119) | fn register_all_operators(&mut self) { method register_projection_operators (line 124) | fn register_projection_operators(&mut self) { function get_operator_type (line 133) | fn get_operator_type(spark_operator: &Operator) -> Option { FILE: native/core/src/execution/serde.rs function deserialize_expr (line 38) | pub fn deserialize_expr(buf: &[u8]) -> Result Result Result Result ArrowDataType { FILE: native/core/src/execution/sort.rs type Rdx (line 23) | pub trait Rdx { method cfg_nbuckets (line 25) | fn cfg_nbuckets() -> usize; method cfg_nrounds (line 28) | fn cfg_nrounds() -> usize; method get_bucket (line 40) | fn get_bucket(&self, round: usize) -> usize; method reverse (line 44) | fn reverse(round: usize, bucket: usize) -> bool; method cfg_nbuckets (line 54) | fn cfg_nbuckets() -> usize { method cfg_nrounds (line 59) | fn cfg_nrounds() -> usize { method get_bucket (line 65) | fn get_bucket(&self, round: usize) -> usize { method reverse (line 73) | fn reverse(_round: usize, _bucket: usize) -> bool { constant MASK_LONG_LOWER_40_BITS (line 47) | const MASK_LONG_LOWER_40_BITS: u64 = (1u64 << 40) - 1; constant MASK_LONG_UPPER_24_BITS (line 48) | const MASK_LONG_UPPER_24_BITS: u64 = !MASK_LONG_LOWER_40_BITS; type RdxSort (line 79) | pub trait RdxSort { method rdxsort (line 81) | fn rdxsort(&mut self); method rdxsort (line 106) | fn rdxsort(&mut self) { function helper_bucket (line 85) | fn helper_bucket(buckets_b: &mut [Vec], iter: I, cfg_nbuckets: ... constant MAXIMUM_PARTITION_ID (line 184) | const MAXIMUM_PARTITION_ID: i32 = (1i32 << 24) - 1; constant MASK_LONG_LOWER_51_BITS (line 185) | const MASK_LONG_LOWER_51_BITS: i64 = (1i64 << 51) - 1; constant MASK_LONG_UPPER_13_BITS (line 186) | const MASK_LONG_UPPER_13_BITS: i64 = !MASK_LONG_LOWER_51_BITS; constant MASK_LONG_LOWER_27_BITS (line 187) | const MASK_LONG_LOWER_27_BITS: i64 = (1i64 << 27) - 1; function pack_pointer (line 190) | fn pack_pointer(pointer: i64, partition_id: i32) -> i64 { function test_rdxsort (line 199) | fn test_rdxsort() { FILE: native/core/src/execution/spark_config.rs constant COMET_TRACING_ENABLED (line 20) | pub(crate) const COMET_TRACING_ENABLED: &str = "spark.comet.tracing.enab... constant COMET_DEBUG_ENABLED (line 21) | pub(crate) const COMET_DEBUG_ENABLED: &str = "spark.comet.debug.enabled"; constant COMET_EXPLAIN_NATIVE_ENABLED (line 22) | pub(crate) const COMET_EXPLAIN_NATIVE_ENABLED: &str = "spark.comet.expla... constant COMET_MAX_TEMP_DIRECTORY_SIZE (line 23) | pub(crate) const COMET_MAX_TEMP_DIRECTORY_SIZE: &str = "spark.comet.maxT... constant COMET_DEBUG_MEMORY (line 24) | pub(crate) const COMET_DEBUG_MEMORY: &str = "spark.comet.debug.memory"; constant SPARK_EXECUTOR_CORES (line 25) | pub(crate) const SPARK_EXECUTOR_CORES: &str = "spark.executor.cores"; type SparkConfig (line 27) | pub(crate) trait SparkConfig { method get_bool (line 28) | fn get_bool(&self, name: &str) -> bool; method get_u64 (line 29) | fn get_u64(&self, name: &str, default_value: u64) -> u64; method get_usize (line 30) | fn get_usize(&self, name: &str, default_value: usize) -> usize; method get_bool (line 34) | fn get_bool(&self, name: &str) -> bool { method get_u64 (line 40) | fn get_u64(&self, name: &str, default_value: u64) -> u64 { method get_usize (line 46) | fn get_usize(&self, name: &str, default_value: usize) -> usize { FILE: native/core/src/execution/spark_plan.rs type SparkPlan (line 26) | pub(crate) struct SparkPlan { method new (line 40) | pub(crate) fn new( method new_with_additional (line 54) | pub(crate) fn new_with_additional( method schema (line 73) | pub(crate) fn schema(&self) -> SchemaRef { method children (line 78) | pub(crate) fn children(&self) -> &Vec> { FILE: native/core/src/execution/utils.rs type SparkArrowConvert (line 25) | pub trait SparkArrowConvert { method from_spark (line 28) | fn from_spark(addresses: (i64, i64)) -> Result method move_to_spark (line 33) | fn move_to_spark(&self, array: i64, schema: i64) -> Result<(), Executi... method from_spark (line 37) | fn from_spark(addresses: (i64, i64)) -> Result { method move_to_spark (line 65) | fn move_to_spark(&self, array: i64, schema: i64) -> Result<(), Executi... FILE: native/core/src/lib.rs function Java_org_apache_comet_NativeBase_init (line 89) | pub extern "system" fn Java_org_apache_comet_NativeBase_init( constant LOG_PATTERN (line 126) | const LOG_PATTERN: &str = "{d(%y/%m/%d %H:%M:%S)} {l} {f}: {m}{n}"; function Java_org_apache_comet_NativeBase_isFeatureEnabled (line 138) | pub extern "system" fn Java_org_apache_comet_NativeBase_isFeatureEnabled( function default_logger_config (line 158) | fn default_logger_config(log_level: &str) -> CometResult { FILE: native/core/src/parquet/cast_column.rs function types_differ_only_in_field_names (line 42) | fn types_differ_only_in_field_names(physical: &DataType, logical: &DataT... function relabel_array (line 80) | fn relabel_array(array: ArrayRef, target_type: &DataType) -> ArrayRef { function cast_timestamp_micros_to_millis_array (line 148) | fn cast_timestamp_micros_to_millis_array( function cast_timestamp_micros_to_millis_scalar (line 172) | fn cast_timestamp_micros_to_millis_scalar( type CometCastColumnExpr (line 181) | pub struct CometCastColumnExpr { method new (line 219) | pub fn new( method with_parquet_options (line 235) | pub fn with_parquet_options(mut self, options: SparkParquetOptions) ->... method eq (line 198) | fn eq(&self, other: &Self) -> bool { method hash (line 208) | fn hash(&self, state: &mut H) { method fmt (line 242) | fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { method as_any (line 253) | fn as_any(&self) -> &dyn Any { method data_type (line 257) | fn data_type(&self, _input_schema: &Schema) -> DataFusionResult { method nullable (line 261) | fn nullable(&self, _input_schema: &Schema) -> DataFusionResult { method evaluate (line 265) | fn evaluate(&self, batch: &RecordBatch) -> DataFusionResult DataFusionResult Vec<&Arc> { method with_new_children (line 331) | fn with_new_children( method fmt_sql (line 349) | fn fmt_sql(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { function test_cast_timestamp_micros_to_millis_array (line 362) | fn test_cast_timestamp_micros_to_millis_array() { function test_cast_timestamp_micros_to_millis_array_with_timezone (line 390) | fn test_cast_timestamp_micros_to_millis_array_with_timezone() { function test_cast_timestamp_micros_to_millis_scalar (line 411) | fn test_cast_timestamp_micros_to_millis_scalar() { function test_comet_cast_column_expr_evaluate_micros_to_millis_array (line 430) | fn test_comet_cast_column_expr_evaluate_micros_to_millis_array() { function test_comet_cast_column_expr_evaluate_micros_to_millis_scalar (line 475) | fn test_comet_cast_column_expr_evaluate_micros_to_millis_scalar() { function test_relabel_list_field_name (line 514) | fn test_relabel_list_field_name() { function test_relabel_map_entries_field_name (line 535) | fn test_relabel_map_entries_field_name() { function test_relabel_struct_metadata (line 571) | fn test_relabel_struct_metadata() { function test_relabel_nested_struct_containing_list (line 593) | fn test_relabel_nested_struct_containing_list() { FILE: native/core/src/parquet/data_type.rs type DataType (line 20) | pub trait DataType: PlainDecoding + PlainDictDecoding + 'static {} type AsBytes (line 60) | pub trait AsBytes { method as_bytes (line 62) | fn as_bytes(&self) -> &[u8]; method as_bytes (line 66) | fn as_bytes(&self) -> &[u8] { method as_bytes (line 72) | fn as_bytes(&self) -> &[u8] { method as_bytes (line 78) | fn as_bytes(&self) -> &[u8] { method as_bytes (line 84) | fn as_bytes(&self) -> &[u8] { FILE: native/core/src/parquet/encryption_support.rs constant ENCRYPTION_FACTORY_ID (line 33) | pub const ENCRYPTION_FACTORY_ID: &str = "comet.jni_kms_encryption"; type CometEncryptionFactory (line 44) | pub struct CometEncryptionFactory { method get_file_encryption_properties (line 52) | async fn get_file_encryption_properties( method get_file_decryption_properties (line 68) | async fn get_file_decryption_properties( type CometKeyRetriever (line 84) | pub struct CometKeyRetriever { method new (line 91) | pub fn new( method retrieve_key (line 115) | fn retrieve_key(&self, key_metadata: &[u8]) -> datafusion::parquet::erro... FILE: native/core/src/parquet/mod.rs type Context (line 74) | struct Context { function Java_org_apache_comet_parquet_Native_initColumnReader (line 79) | pub extern "system" fn Java_org_apache_comet_parquet_Native_initColumnRe... function Java_org_apache_comet_parquet_Native_setDictionaryPage (line 141) | pub unsafe extern "system" fn Java_org_apache_comet_parquet_Native_setDi... function Java_org_apache_comet_parquet_Native_setPageV1 (line 168) | pub unsafe extern "system" fn Java_org_apache_comet_parquet_Native_setPa... function Java_org_apache_comet_parquet_Native_setPageV2 (line 195) | pub unsafe extern "system" fn Java_org_apache_comet_parquet_Native_setPa... function Java_org_apache_comet_parquet_Native_resetBatch (line 236) | pub extern "system" fn Java_org_apache_comet_parquet_Native_resetBatch( function Java_org_apache_comet_parquet_Native_readBatch (line 249) | pub extern "system" fn Java_org_apache_comet_parquet_Native_readBatch( function Java_org_apache_comet_parquet_Native_skipBatch (line 268) | pub extern "system" fn Java_org_apache_comet_parquet_Native_skipBatch( function Java_org_apache_comet_parquet_Native_currentBatch (line 282) | pub extern "system" fn Java_org_apache_comet_parquet_Native_currentBatch( function get_context (line 299) | fn get_context<'a>(handle: jlong) -> Result<&'a mut Context, CometError> { function get_reader (line 308) | fn get_reader<'a>(handle: jlong) -> Result<&'a mut ColumnReader, CometEr... function Java_org_apache_comet_parquet_Native_closeColumnReader (line 313) | pub extern "system" fn Java_org_apache_comet_parquet_Native_closeColumnR... function from_u8_slice (line 327) | fn from_u8_slice(src: &mut [u8]) -> &mut [i8] { type ParquetReaderState (line 333) | enum ParquetReaderState { type BatchContext (line 339) | struct BatchContext { function get_batch_context (line 348) | fn get_batch_context<'a>(handle: jlong) -> Result<&'a mut BatchContext, ... function get_file_groups_single_file (line 356) | fn get_file_groups_single_file( function get_object_store_options (line 377) | pub fn get_object_store_options( function Java_org_apache_comet_parquet_Native_validateObjectStoreConfig (line 403) | pub unsafe extern "system" fn Java_org_apache_comet_parquet_Native_valid... function Java_org_apache_comet_parquet_Native_initRecordBatchReader (line 428) | pub unsafe extern "system" fn Java_org_apache_comet_parquet_Native_initR... function Java_org_apache_comet_parquet_Native_readNextRecordBatch (line 535) | pub extern "system" fn Java_org_apache_comet_parquet_Native_readNextReco... function Java_org_apache_comet_parquet_Native_currentColumnBatch (line 580) | pub extern "system" fn Java_org_apache_comet_parquet_Native_currentColum... function Java_org_apache_comet_parquet_Native_closeRecordBatchReader (line 603) | pub extern "system" fn Java_org_apache_comet_parquet_Native_closeRecordB... FILE: native/core/src/parquet/mutable_vector.rs constant DEFAULT_ARRAY_LEN (line 23) | const DEFAULT_ARRAY_LEN: usize = 4; type ParquetMutableVector (line 34) | pub struct ParquetMutableVector { method new (line 62) | pub fn new(capacity: usize, arrow_type: &ArrowDataType) -> Self { method new_with_bit_width (line 67) | pub fn new_with_bit_width( method is_null (line 116) | pub fn is_null(&self, idx: usize) -> bool { method reset (line 122) | pub fn reset(&mut self) { method put_null (line 140) | pub fn put_null(&mut self) { method put_nulls (line 146) | pub fn put_nulls(&mut self, n: usize) { method num_values (line 164) | pub fn num_values(&self) -> usize { method num_nulls (line 170) | pub fn num_nulls(&self) -> usize { method set_dictionary (line 175) | pub fn set_dictionary(&mut self, dict: ParquetMutableVector) { method get_array_data (line 188) | pub fn get_array_data(&mut self) -> Result { method bit_width (line 218) | pub fn bit_width(arrow_type: &ArrowDataType) -> usize { method is_binary_type (line 233) | fn is_binary_type(dt: &ArrowDataType) -> bool { method should_reset_value_buffer (line 238) | fn should_reset_value_buffer(dt: &ArrowDataType) -> bool { FILE: native/core/src/parquet/objectstore/s3.rs function create_store (line 58) | pub fn create_store( function region_cache (line 136) | fn region_cache() -> &'static RwLock> { function resolve_bucket_region (line 148) | pub async fn resolve_bucket_region(bucket: &str) -> Result( function get_config_trimmed (line 289) | fn get_config_trimmed<'a>( constant HADOOP_IAM_INSTANCE (line 298) | const HADOOP_IAM_INSTANCE: &str = "org.apache.hadoop.fs.s3a.auth.IAMInst... constant HADOOP_SIMPLE (line 299) | const HADOOP_SIMPLE: &str = "org.apache.hadoop.fs.s3a.SimpleAWSCredentia... constant HADOOP_TEMPORARY (line 300) | const HADOOP_TEMPORARY: &str = "org.apache.hadoop.fs.s3a.TemporaryAWSCre... constant HADOOP_ASSUMED_ROLE (line 301) | const HADOOP_ASSUMED_ROLE: &str = "org.apache.hadoop.fs.s3a.auth.Assumed... constant HADOOP_ANONYMOUS (line 302) | const HADOOP_ANONYMOUS: &str = "org.apache.hadoop.fs.s3a.AnonymousAWSCre... constant AWS_CONTAINER_CREDENTIALS (line 305) | const AWS_CONTAINER_CREDENTIALS: &str = constant AWS_CONTAINER_CREDENTIALS_V1 (line 307) | const AWS_CONTAINER_CREDENTIALS_V1: &str = "com.amazonaws.auth.Container... constant AWS_EC2_CONTAINER_CREDENTIALS (line 308) | const AWS_EC2_CONTAINER_CREDENTIALS: &str = constant AWS_INSTANCE_PROFILE (line 310) | const AWS_INSTANCE_PROFILE: &str = constant AWS_INSTANCE_PROFILE_V1 (line 312) | const AWS_INSTANCE_PROFILE_V1: &str = "com.amazonaws.auth.InstanceProfil... constant AWS_ENVIRONMENT (line 313) | const AWS_ENVIRONMENT: &str = constant AWS_ENVIRONMENT_V1 (line 315) | const AWS_ENVIRONMENT_V1: &str = "com.amazonaws.auth.EnvironmentVariable... constant AWS_WEB_IDENTITY (line 316) | const AWS_WEB_IDENTITY: &str = constant AWS_WEB_IDENTITY_V1 (line 318) | const AWS_WEB_IDENTITY_V1: &str = "com.amazonaws.auth.WebIdentityTokenCr... constant AWS_ANONYMOUS (line 319) | const AWS_ANONYMOUS: &str = "software.amazon.awssdk.auth.credentials.Ano... constant AWS_ANONYMOUS_V1 (line 320) | const AWS_ANONYMOUS_V1: &str = "com.amazonaws.auth.AnonymousAWSCredentia... function build_credential_provider (line 337) | async fn build_credential_provider( function parse_credential_provider_names (line 378) | fn parse_credential_provider_names(aws_credential_provider_names: &str) ... function is_anonymous_credential_provider (line 386) | fn is_anonymous_credential_provider(credential_provider_name: &str) -> b... function build_chained_aws_credential_provider_metadata (line 390) | fn build_chained_aws_credential_provider_metadata( function build_aws_credential_provider_metadata (line 422) | fn build_aws_credential_provider_metadata( function build_static_credential_provider_metadata (line 449) | fn build_static_credential_provider_metadata( function build_assume_role_credential_provider_metadata (line 479) | fn build_assume_role_credential_provider_metadata( type CachedAwsCredentialProvider (line 529) | struct CachedAwsCredentialProvider { method new (line 552) | fn new( method metadata (line 567) | fn metadata(&self) -> CredentialProviderMetadata { method fetch_credential (line 571) | fn fetch_credential(&self) -> Option { method refresh_credential (line 585) | async fn refresh_credential(&self) -> object_store::Result object_store::Result) ... method provide_credentials (line 644) | fn provide_credentials<'a>( type CredentialProviderMetadata (line 666) | enum CredentialProviderMetadata { method name (line 687) | fn name(&self) -> &'static str { method simple_string (line 702) | fn simple_string(&self) -> String { method create_credential_provider (line 738) | async fn create_credential_provider( type TestConfigBuilder (line 830) | struct TestConfigBuilder { method new (line 835) | fn new() -> Self { method with_region (line 839) | fn with_region(mut self, region: &str) -> Self { method with_credential_provider (line 845) | fn with_credential_provider(mut self, provider: &str) -> Self { method with_bucket_credential_provider (line 853) | fn with_bucket_credential_provider(mut self, bucket: &str, provider: &... method with_access_key (line 861) | fn with_access_key(mut self, key: &str) -> Self { method with_secret_key (line 867) | fn with_secret_key(mut self, key: &str) -> Self { method with_session_token (line 873) | fn with_session_token(mut self, token: &str) -> Self { method with_bucket_access_key (line 879) | fn with_bucket_access_key(mut self, bucket: &str, key: &str) -> Self { method with_bucket_secret_key (line 887) | fn with_bucket_secret_key(mut self, bucket: &str, key: &str) -> Self { method with_bucket_session_token (line 895) | fn with_bucket_session_token(mut self, bucket: &str, token: &str) -> S... method with_assume_role_arn (line 903) | fn with_assume_role_arn(mut self, arn: &str) -> Self { method with_assume_role_session_name (line 909) | fn with_assume_role_session_name(mut self, name: &str) -> Self { method with_assume_role_credentials_provider (line 917) | fn with_assume_role_credentials_provider(mut self, provider: &str) -> ... method build (line 925) | fn build(self) -> HashMap { function test_create_store (line 932) | fn test_create_store() { function test_get_config_trimmed (line 947) | fn test_get_config_trimmed() { function test_parse_credential_provider_names (line 975) | fn test_parse_credential_provider_names() { function test_default_credential_provider (line 1013) | async fn test_default_credential_provider() { function test_anonymous_credential_provider (line 1040) | async fn test_anonymous_credential_provider() { function test_mixed_anonymous_and_other_providers_error (line 1056) | async fn test_mixed_anonymous_and_other_providers_error() { function test_simple_credential_provider (line 1077) | async fn test_simple_credential_provider() { function test_temporary_credential_provider (line 1106) | async fn test_temporary_credential_provider() { function test_missing_access_key (line 1135) | async fn test_missing_access_key() { function test_missing_secret_key (line 1161) | async fn test_missing_secret_key() { function test_missing_session_token_for_temporary (line 1187) | async fn test_missing_session_token_for_temporary() { function test_bucket_specific_configuration (line 1214) | async fn test_bucket_specific_configuration() { function test_assume_role_credential_provider (line 1300) | async fn test_assume_role_credential_provider() { function test_assume_role_missing_arn_error (line 1336) | async fn test_assume_role_missing_arn_error() { function test_unsupported_credential_provider_error (line 1353) | async fn test_unsupported_credential_provider_error() { function test_environment_credential_provider (line 1372) | async fn test_environment_credential_provider() { function test_ecs_credential_provider (line 1391) | async fn test_ecs_credential_provider() { function test_imds_credential_provider (line 1414) | async fn test_imds_credential_provider() { function test_web_identity_credential_provider (line 1433) | async fn test_web_identity_credential_provider() { function test_hadoop_iam_instance_credential_provider (line 1452) | async fn test_hadoop_iam_instance_credential_provider() { function test_chained_credential_providers (line 1474) | async fn test_chained_credential_providers() { function test_static_environment_web_identity_chain (line 1502) | async fn test_static_environment_web_identity_chain() { function test_assume_role_with_static_base_provider (line 1537) | async fn test_assume_role_with_static_base_provider() { function test_assume_role_with_web_identity_base_provider (line 1573) | async fn test_assume_role_with_web_identity_base_provider() { function test_assume_role_with_chained_base_providers (line 1601) | async fn test_assume_role_with_chained_base_providers() { function test_assume_role_chained_with_other_providers (line 1643) | async fn test_assume_role_chained_with_other_providers() { function test_assume_role_with_anonymous_base_provider_error (line 1691) | async fn test_assume_role_with_anonymous_base_provider_error() { function test_get_credential_from_static_credential_provider (line 1716) | async fn test_get_credential_from_static_credential_provider() { function test_get_credential_from_invalid_static_credential_provider (line 1754) | async fn test_get_credential_from_invalid_static_credential_provider() { function test_invalid_static_credential_provider_should_not_prevent_other_providers_from_working (line 1772) | async fn test_invalid_static_credential_provider_should_not_prevent_othe... type MockAwsCredentialProvider (line 1813) | struct MockAwsCredentialProvider { method provide_credentials (line 1818) | fn provide_credentials<'a>( function test_cached_credential_provider_refresh_credential (line 1837) | async fn test_cached_credential_provider_refresh_credential() { function test_cached_credential_provider_cache_credential (line 1857) | async fn test_cached_credential_provider_cache_credential() { function test_extract_s3_config_options (line 1876) | fn test_extract_s3_config_options() { function test_extract_s3_config_custom_endpoint (line 1898) | fn test_extract_s3_config_custom_endpoint() { function test_extract_s3_config_custom_endpoint_with_virtual_hosted_style (line 1919) | fn test_extract_s3_config_custom_endpoint_with_virtual_hosted_style() { function test_extract_s3_config_ignore_default_endpoint (line 1955) | fn test_extract_s3_config_ignore_default_endpoint() { function test_credential_provider_metadata_simple_string (line 1970) | fn test_credential_provider_metadata_simple_string() { FILE: native/core/src/parquet/parquet_exec.rs function init_datasource_exec (line 63) | pub(crate) fn init_datasource_exec( function get_options (line 185) | fn get_options( function dbg_batch_stream (line 214) | pub fn dbg_batch_stream(stream: SendableRecordBatchStream) -> SendableRe... FILE: native/core/src/parquet/parquet_read_cached_factory.rs type MetadataCell (line 47) | type MetadataCell = Arc>>; type CachingParquetReaderFactory (line 53) | pub struct CachingParquetReaderFactory { method new (line 59) | pub fn new(store: Arc) -> Self { method create_reader (line 68) | fn create_reader( type CachingParquetFileReader (line 104) | struct CachingParquetFileReader { method get_bytes (line 112) | fn get_bytes(&mut self, range: Range) -> BoxFuture<'_, parquet::err... method get_byte_ranges (line 117) | fn get_byte_ranges( method get_metadata (line 129) | fn get_metadata<'a>( FILE: native/core/src/parquet/parquet_support.rs type SparkParquetOptions (line 65) | pub struct SparkParquetOptions { method new (line 85) | pub fn new(eval_mode: EvalMode, timezone: &str, allow_incompat: bool) ... method new_without_timezone (line 97) | pub fn new_without_timezone(eval_mode: EvalMode, allow_incompat: bool)... function spark_parquet_convert (line 113) | pub fn spark_parquet_convert( function parquet_convert_array (line 138) | fn parquet_convert_array( function parquet_convert_struct_to_struct (line 238) | fn parquet_convert_struct_to_struct( function parquet_convert_map_to_map (line 302) | fn parquet_convert_map_to_map( function key_field (line 347) | fn key_field(entries_field: &FieldRef) -> Option { function value_field (line 356) | fn value_field(entries_field: &FieldRef) -> Option { function is_hdfs_scheme (line 364) | pub fn is_hdfs_scheme(url: &Url, object_store_configs: &HashMap Result Result { function create_hdfs_object_store (line 441) | fn create_hdfs_object_store( type ObjectStoreCache (line 450) | type ObjectStoreCache = RwLock &'static ObjectStoreCache { function hash_object_store_configs (line 489) | fn hash_object_store_configs(configs: &HashMap) -> u64 { function prepare_object_store_with_configs (line 502) | pub(crate) fn prepare_object_store_with_configs( function prepare_object_store (line 600) | pub(crate) fn prepare_object_store( function prepare_object_store (line 610) | pub(crate) fn prepare_object_store( function test_prepare_object_store (line 621) | fn test_prepare_object_store() { function test_prepare_object_store (line 664) | fn test_prepare_object_store() { FILE: native/core/src/parquet/read/column.rs constant DECIMAL_MAX_INT_DIGITS (line 43) | const DECIMAL_MAX_INT_DIGITS: i32 = 9; constant DECIMAL_MAX_LONG_DIGITS (line 46) | const DECIMAL_MAX_LONG_DIGITS: i32 = 18; type ColumnReader (line 48) | pub enum ColumnReader { method get (line 93) | pub fn get( method get_descriptor (line 531) | pub fn get_descriptor(&self) -> &ColumnDescriptor { method set_dictionary_page (line 536) | pub fn set_dictionary_page( method set_page_v1 (line 552) | pub fn set_page_v1(&mut self, page_value_count: usize, page_data: Buff... method set_page_v2 (line 557) | pub fn set_page_v2( method reset_batch (line 577) | pub fn reset_batch(&mut self) { method current_batch (line 582) | pub fn current_batch(&mut self) -> Result { method read_batch (line 587) | pub fn read_batch(&mut self, total: usize, null_pad_size: usize) -> (u... method skip_batch (line 592) | pub fn skip_batch(&mut self, total: usize, put_nulls: bool) -> usize { type TypedColumnReader (line 598) | pub struct TypedColumnReader { function new (line 621) | pub fn new( function get_descriptor (line 645) | pub fn get_descriptor(&self) -> &ColumnDescriptor { function reset_batch (line 652) | pub fn reset_batch(&mut self) { function current_batch (line 661) | pub fn current_batch(&mut self) -> Result { function read_batch (line 673) | pub fn read_batch(&mut self, total: usize, null_pad_size: usize) -> (usi... function skip_batch (line 696) | pub fn skip_batch(&mut self, total: usize, put_nulls: bool) -> usize { function set_dictionary_page (line 719) | pub fn set_dictionary_page( function set_page_v1 (line 758) | pub fn set_page_v1( function set_page_v2 (line 794) | pub fn set_page_v2( function check_dictionary (line 819) | fn check_dictionary(&mut self, encoding: &Encoding) { function get_decoder (line 831) | fn get_decoder(&self, value_data: Buffer, encoding: Encoding) -> Box S... method set_data (line 81) | pub fn set_data(&mut self, page_value_count: usize, page_data: &Buffer... method read_batch (line 103) | pub fn read_batch( method skip_batch (line 151) | pub fn skip_batch( method read_next_group (line 197) | fn read_next_group(&mut self) { FILE: native/core/src/parquet/read/mod.rs type ReadOptions (line 31) | pub struct ReadOptions { type PlainDecoderInner (line 37) | pub struct PlainDecoderInner { type PlainDecoding (line 58) | pub trait PlainDecoding { method decode (line 63) | fn decode(src: &mut PlainDecoderInner, dst: &mut ParquetMutableVector,... method skip (line 69) | fn skip(src: &mut PlainDecoderInner, num: usize); type PlainDictDecoding (line 72) | pub trait PlainDictDecoding { method decode_dict (line 75) | fn decode_dict(src: ParquetMutableVector, dst: &mut ParquetMutableVect... method decode_dict_one (line 101) | fn decode_dict_one( FILE: native/core/src/parquet/read/values.rs function get_decoder (line 35) | pub fn get_decoder( type Decoder (line 53) | pub trait Decoder { method read (line 60) | fn read(&mut self, dst: &mut ParquetMutableVector); method read_batch (line 68) | fn read_batch(&mut self, dst: &mut ParquetMutableVector, num: usize); method skip_batch (line 75) | fn skip_batch(&mut self, num: usize); method encoding (line 78) | fn encoding(&self) -> Encoding; method read (line 899) | fn read(&mut self, dst: &mut ParquetMutableVector) { method read_batch (line 908) | fn read_batch(&mut self, dst: &mut ParquetMutableVector, num: usize) { method skip_batch (line 913) | fn skip_batch(&mut self, num: usize) { method encoding (line 918) | fn encoding(&self) -> Encoding { method read (line 975) | fn read(&mut self, dst: &mut ParquetMutableVector) { method read_batch (line 999) | fn read_batch(&mut self, dst: &mut ParquetMutableVector, num: usize) { method skip_batch (line 1036) | fn skip_batch(&mut self, num: usize) { method encoding (line 1057) | fn encoding(&self) -> Encoding { constant JULIAN_GREGORIAN_SWITCH_OFF_DAY (line 83) | const JULIAN_GREGORIAN_SWITCH_OFF_DAY: i32 = -141427; constant JULIAN_GREGORIAN_SWITCH_OFF_TS (line 87) | const JULIAN_GREGORIAN_SWITCH_OFF_TS: i64 = -2208988800000000; constant JULIAN_DAY_OF_EPOCH (line 91) | const JULIAN_DAY_OF_EPOCH: i32 = 2440588; constant MICROS_PER_MILLIS (line 94) | const MICROS_PER_MILLIS: i64 = 1000; constant MICROS_PER_DAY (line 96) | const MICROS_PER_DAY: i64 = 24_i64 * 60 * 60 * 1000 * 1000; type PlainDecoder (line 98) | pub struct PlainDecoder { function new (line 107) | pub fn new(value_data: Buffer, desc: ColumnDescPtr, read_options: ReadOp... method decode (line 212) | fn decode(src: &mut PlainDecoderInner, dst: &mut ParquetMutableVector, n... method skip (line 252) | fn skip(src: &mut PlainDecoderInner, num: usize) { method decode (line 260) | fn decode(src: &mut PlainDecoderInner, dst: &mut ParquetMutableVector, n... method skip (line 303) | fn skip(src: &mut PlainDecoderInner, num: usize) { method decode (line 311) | fn decode(src: &mut PlainDecoderInner, dst: &mut ParquetMutableVector, n... method skip (line 357) | fn skip(src: &mut PlainDecoderInner, num: usize) { method decode (line 365) | fn decode(src: &mut PlainDecoderInner, dst: &mut ParquetMutableVector, n... method skip (line 404) | fn skip(src: &mut PlainDecoderInner, num: usize) { method decode (line 415) | fn decode(src: &mut PlainDecoderInner, dst: &mut ParquetMutableVector, n... method skip (line 421) | fn skip(src: &mut PlainDecoderInner, num: usize) { method decode_dict_one (line 429) | fn decode_dict_one( constant INT96_SRC_BYTE_WIDTH (line 815) | const INT96_SRC_BYTE_WIDTH: usize = 12; constant INT96_DST_BYTE_WIDTH (line 817) | const INT96_DST_BYTE_WIDTH: usize = 8; function int96_to_microsecond (line 819) | fn int96_to_microsecond(v: &[u8]) -> i64 { method decode (line 833) | fn decode(src: &mut PlainDecoderInner, dst: &mut ParquetMutableVector, n... method skip (line 874) | fn skip(src: &mut PlainDecoderInner, num: usize) { method decode_dict_one (line 880) | fn decode_dict_one( type DictDecoder (line 925) | pub struct DictDecoder { method new (line 943) | pub fn new(buf: Buffer) -> Self { method reload (line 959) | fn reload(&mut self) { function test_i32_to_i8 (line 1068) | fn test_i32_to_i8() { function test_i32_to_u8 (line 1080) | fn test_i32_to_u8() { function test_i32_to_i16 (line 1092) | fn test_i32_to_i16() { function test_i32_to_u16 (line 1104) | fn test_i32_to_u16() { function test_i32_to_u32 (line 1120) | fn test_i32_to_u32() { FILE: native/core/src/parquet/schema_adapter.rs type SparkPhysicalExprAdapterFactory (line 41) | pub struct SparkPhysicalExprAdapterFactory { method new (line 51) | pub fn new( function remap_physical_schema_names (line 66) | fn remap_physical_schema_names( function check_column_duplicate (line 99) | fn check_column_duplicate(col_name: &str, physical_schema: &SchemaRef) -... method create (line 115) | fn create( type SparkPhysicalExprAdapter (line 187) | struct SparkPhysicalExprAdapter { method wrap_all_type_mismatches (line 293) | fn wrap_all_type_mismatches( method replace_with_spark_cast (line 374) | fn replace_with_spark_cast( method replace_missing_with_defaults (line 449) | fn replace_missing_with_defaults( method rewrite (line 211) | fn rewrite(&self, expr: Arc) -> DataFusionResult Result<(), DataFusionError> { function parquet_roundtrip_unsigned_int (line 568) | async fn parquet_roundtrip_unsigned_int() -> Result<(), DataFusionError> { function roundtrip (line 583) | async fn roundtrip( function parquet_duplicate_fields_case_insensitive (line 620) | async fn parquet_duplicate_fields_case_insensitive() { FILE: native/core/src/parquet/util/bit_packing.rs function unpack32 (line 29) | pub unsafe fn unpack32(mut in_ptr: *const u32, out_ptr: *mut u32, num_bi... function nullunpacker32 (line 69) | unsafe fn nullunpacker32(in_buf: *const u32, mut out: *mut u32) -> *cons... function unpack1_32 (line 77) | unsafe fn unpack1_32(in_buf: *const u32, mut out: *mut u32) -> *const u32 { function unpack2_32 (line 145) | unsafe fn unpack2_32(mut in_buf: *const u32, mut out: *mut u32) -> *cons... function unpack3_32 (line 214) | unsafe fn unpack3_32(mut in_buf: *const u32, mut out: *mut u32) -> *cons... function unpack4_32 (line 288) | unsafe fn unpack4_32(mut in_buf: *const u32, mut out: *mut u32) -> *cons... function unpack5_32 (line 362) | unsafe fn unpack5_32(mut in_buf: *const u32, mut out: *mut u32) -> *cons... function unpack6_32 (line 442) | unsafe fn unpack6_32(mut in_buf: *const u32, mut out: *mut u32) -> *cons... function unpack7_32 (line 524) | unsafe fn unpack7_32(mut in_buf: *const u32, mut out: *mut u32) -> *cons... function unpack8_32 (line 610) | unsafe fn unpack8_32(mut in_buf: *const u32, mut out: *mut u32) -> *cons... function unpack9_32 (line 692) | unsafe fn unpack9_32(mut in_buf: *const u32, mut out: *mut u32) -> *cons... function unpack10_32 (line 784) | unsafe fn unpack10_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack11_32 (line 878) | unsafe fn unpack11_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack12_32 (line 976) | unsafe fn unpack12_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack13_32 (line 1074) | unsafe fn unpack13_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack14_32 (line 1178) | unsafe fn unpack14_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack15_32 (line 1284) | unsafe fn unpack15_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack16_32 (line 1394) | unsafe fn unpack16_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack17_32 (line 1492) | unsafe fn unpack17_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack18_32 (line 1608) | unsafe fn unpack18_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack19_32 (line 1726) | unsafe fn unpack19_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack20_32 (line 1848) | unsafe fn unpack20_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack21_32 (line 1970) | unsafe fn unpack21_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack22_32 (line 2098) | unsafe fn unpack22_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack23_32 (line 2228) | unsafe fn unpack23_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack24_32 (line 2362) | unsafe fn unpack24_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack25_32 (line 2492) | unsafe fn unpack25_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack26_32 (line 2632) | unsafe fn unpack26_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack27_32 (line 2774) | unsafe fn unpack27_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack28_32 (line 2920) | unsafe fn unpack28_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack29_32 (line 3066) | unsafe fn unpack29_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack30_32 (line 3218) | unsafe fn unpack30_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack31_32 (line 3372) | unsafe fn unpack31_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... function unpack32_32 (line 3530) | unsafe fn unpack32_32(mut in_buf: *const u32, mut out: *mut u32) -> *con... FILE: native/core/src/parquet/util/buffer.rs type Buffer (line 21) | pub trait Buffer { method len (line 23) | fn len(&self) -> usize; method data (line 26) | fn data(&self) -> &[u8]; method is_empty (line 29) | fn is_empty(&self) -> bool { method len (line 35) | fn len(&self) -> usize { method data (line 39) | fn data(&self) -> &[u8] { type BufferRef (line 44) | pub struct BufferRef { method new (line 51) | pub fn new(inner: Arc) -> Self { method len (line 62) | pub fn len(&self) -> usize { method is_empty (line 67) | pub fn is_empty(&self) -> bool { method data (line 72) | pub fn data(&self) -> &[u8] { method slice (line 78) | pub fn slice(&self, offset: usize, len: usize) -> BufferRef { method start (line 98) | pub fn start(&self, offset: usize) -> BufferRef { method as_ref (line 113) | fn as_ref(&self) -> &[u8] { type Output (line 123) | type Output = Idx::Output; method index (line 125) | fn index(&self, index: Idx) -> &Self::Output { FILE: native/core/src/parquet/util/jni.rs function convert_column_descriptor (line 39) | pub fn convert_column_descriptor( function convert_encoding (line 85) | pub fn convert_encoding(ordinal: jint) -> Encoding { type TypePromotionInfo (line 101) | pub struct TypePromotionInfo { method new_from_jni (line 109) | pub fn new_from_jni( method new (line 124) | pub fn new(physical_type: PhysicalType, precision: i32, scale: i32, bi... function convert_column_path (line 134) | fn convert_column_path(env: &mut Env, path_array: JObjectArray) -> JNIRe... function convert_physical_type (line 145) | fn convert_physical_type(id: jint) -> PhysicalType { function convert_logical_type (line 159) | fn convert_logical_type( function convert_time_unit (line 186) | fn convert_time_unit(time_unit: jint) -> TimeUnit { function fix_type_length (line 197) | fn fix_type_length(t: &PhysicalType, type_length: i32) -> i32 { function deserialize_schema (line 206) | pub fn deserialize_schema(ipc_bytes: &[u8]) -> Result Result<(ObjectStoreUrl, Path), Par... function test_get_file_path (line 246) | fn test_get_file_path() { FILE: native/core/src/parquet/util/memory.rs type MemTrackerPtr (line 35) | pub type MemTrackerPtr = Arc; type WeakMemTrackerPtr (line 37) | pub type WeakMemTrackerPtr = Weak; type MemTracker (line 41) | pub struct MemTracker { method new (line 51) | pub fn new() -> MemTracker { method memory_usage (line 59) | pub fn memory_usage(&self) -> i64 { method max_memory_usage (line 64) | pub fn max_memory_usage(&self) -> i64 { method alloc (line 70) | pub fn alloc(&self, num_bytes: i64) { method default (line 81) | fn default() -> Self { type ByteBuffer (line 90) | pub type ByteBuffer = Buffer; type ByteBufferPtr (line 92) | pub type ByteBufferPtr = BufferPtr; type Buffer (line 101) | pub struct Buffer { function new (line 109) | pub fn new() -> Self { function with_mem_tracker (line 119) | pub fn with_mem_tracker(mut self, mc: MemTrackerPtr) -> Self { function data (line 127) | pub fn data(&self) -> &[T] { function set_data (line 133) | pub fn set_data(&mut self, new_data: Vec) { function resize (line 148) | pub fn resize(&mut self, new_size: usize, init_value: T) { function clear (line 159) | pub fn clear(&mut self) { function reserve (line 167) | pub fn reserve(&mut self, additional_capacity: usize) { function consume (line 181) | pub fn consume(&mut self) -> BufferPtr { function push (line 192) | pub fn push(&mut self, value: T) { function capacity (line 198) | pub fn capacity(&self) -> usize { function size (line 204) | pub fn size(&self) -> usize { function is_mem_tracked (line 210) | pub fn is_mem_tracked(&self) -> bool { function mem_tracker (line 218) | pub fn mem_tracker(&self) -> &MemTrackerPtr { method default (line 224) | fn default() -> Self { type Output (line 230) | type Output = T; function index (line 232) | fn index(&self, index: usize) -> &T { function index_mut (line 238) | fn index_mut(&mut self, index: usize) -> &mut T { method write (line 246) | fn write(&mut self, buf: &[u8]) -> IoResult { method flush (line 257) | fn flush(&mut self) -> IoResult<()> { function as_ref (line 264) | fn as_ref(&self) -> &[u8] { method drop (line 271) | fn drop(&mut self) { type BufferPtr (line 286) | pub struct BufferPtr { function new (line 296) | pub fn new(v: Vec) -> Self { function data (line 308) | pub fn data(&self) -> &[T] { function with_range (line 316) | pub fn with_range(mut self, start: usize, len: usize) -> Self { function set_range (line 325) | pub fn set_range(&mut self, start: usize, len: usize) { function with_mem_tracker (line 332) | pub fn with_mem_tracker(mut self, mc: MemTrackerPtr) -> Self { function start (line 339) | pub fn start(&self) -> usize { function len (line 345) | pub fn len(&self) -> usize { function is_empty (line 351) | pub fn is_empty(&self) -> bool { function is_mem_tracked (line 356) | pub fn is_mem_tracked(&self) -> bool { function all (line 362) | pub fn all(&self) -> BufferPtr { function start_from (line 372) | pub fn start_from(&self, start: usize) -> BufferPtr { function range (line 383) | pub fn range(&self, start: usize, len: usize) -> BufferPtr { type Output (line 395) | type Output = T; function index (line 397) | fn index(&self, index: usize) -> &T { method fmt (line 404) | fn fmt(&self, f: &mut Formatter) -> FmtResult { method drop (line 410) | fn drop(&mut self) { function as_ref (line 421) | fn as_ref(&self) -> &[u8] { function test_byte_buffer_mem_tracker (line 431) | fn test_byte_buffer_mem_tracker() { function test_byte_ptr_mem_tracker (line 467) | fn test_byte_ptr_mem_tracker() { function test_byte_buffer (line 493) | fn test_byte_buffer() { function test_byte_ptr (line 532) | fn test_byte_ptr() { FILE: native/core/src/parquet/util/test_common/page_util.rs type DataPageBuilder (line 37) | pub trait DataPageBuilder { method add_rep_levels (line 38) | fn add_rep_levels(&mut self, max_level: i16, rep_levels: &[i16]); method add_def_levels (line 39) | fn add_def_levels(&mut self, max_level: i16, def_levels: &[i16]); method add_values (line 40) | fn add_values(&mut self, encoding: Encoding, values: &[T:... method add_indices (line 41) | fn add_indices(&mut self, indices: Bytes); method consume (line 42) | fn consume(self) -> Page; method add_rep_levels (line 101) | fn add_rep_levels(&mut self, max_levels: i16, rep_levels: &[i16]) { method add_def_levels (line 106) | fn add_def_levels(&mut self, max_levels: i16, def_levels: &[i16]) { method add_values (line 115) | fn add_values(&mut self, encoding: Encoding, values: &[T:... method add_indices (line 132) | fn add_indices(&mut self, indices: Bytes) { method consume (line 137) | fn consume(self) -> Page { type DataPageBuilderImpl (line 51) | pub struct DataPageBuilderImpl { method new (line 65) | pub fn new(desc: ColumnDescPtr, num_values: u32, datapage_v2: bool) ->... method add_levels (line 78) | fn add_levels(&mut self, max_level: i16, levels: &[i16]) -> u32 { type InMemoryPageReader (line 166) | pub struct InMemoryPageReader> { function new (line 171) | pub fn new(pages: impl IntoIterator) -> Self { method get_next_page (line 179) | fn get_next_page(&mut self) -> Result> { method peek_next_page (line 183) | fn peek_next_page(&mut self) -> Result> { method skip_next_page (line 187) | fn skip_next_page(&mut self) -> Result<()> { type Item (line 193) | type Item = Result; method next (line 195) | fn next(&mut self) -> Option { type InMemoryPageIterator (line 202) | pub struct InMemoryPageIterator>> { function new (line 207) | pub fn new(pages: impl IntoIterator, IntoIter = I>) -> ... type Item (line 215) | type Item = Result>; method next (line 217) | fn next(&mut self) -> Option { function make_pages (line 227) | pub fn make_pages( FILE: native/core/src/parquet/util/test_common/rand_gen.rs function random_bytes (line 23) | pub fn random_bytes(n: usize) -> Vec { function random_bools (line 32) | pub fn random_bools(n: usize) -> Vec { function random_numbers (line 41) | pub fn random_numbers(n: usize) -> Vec function random_numbers_range (line 49) | pub fn random_numbers_range(n: usize, low: T, high: T, result: &mut V... FILE: native/fs-hdfs/build.rs function main (line 21) | fn main() { function build_ffi (line 31) | fn build_ffi(flags: &[String]) { function build_hdfs_lib (line 60) | fn build_hdfs_lib(flags: &[String]) { function build_minidfs_lib (line 89) | fn build_minidfs_lib(flags: &[String]) { function get_build_flags (line 111) | fn get_build_flags() -> Vec { function get_java_dependency (line 120) | fn get_java_dependency() -> Vec { function get_hdfs_file_path (line 144) | fn get_hdfs_file_path(filename: &'static str) -> String { function get_hdfs_file_os_path (line 148) | fn get_hdfs_file_os_path(filename: &'static str) -> String { function get_hdfs_source_dir (line 152) | fn get_hdfs_source_dir() -> &'static str { function get_minidfs_file_path (line 156) | fn get_minidfs_file_path(filename: &'static str) -> String { FILE: native/fs-hdfs/c_src/libhdfs/exception.c type ExceptionInfo (line 30) | struct ExceptionInfo { type ExceptionInfo (line 36) | struct ExceptionInfo function getExceptionInfo (line 94) | void getExceptionInfo(const char *excName, int noPrintFlags, function printExceptionAndFreeV (line 149) | int printExceptionAndFreeV(JNIEnv *env, jthrowable exc, int noPrintFlags, function printExceptionAndFree (line 206) | int printExceptionAndFree(JNIEnv *env, jthrowable exc, int noPrintFlags, function printPendingExceptionAndFree (line 218) | int printPendingExceptionAndFree(JNIEnv *env, int noPrintFlags, function jthrowable (line 241) | jthrowable getPendingExceptionAndClear(JNIEnv *env) function jthrowable (line 250) | jthrowable newRuntimeError(JNIEnv *env, const char *fmt, ...) FILE: native/fs-hdfs/c_src/libhdfs/hdfs.c type hdfsStreamType (line 67) | enum hdfsStreamType type hdfsFile_internal (line 77) | struct hdfsFile_internal { type hdfsExtendedFileInfo (line 88) | struct hdfsExtendedFileInfo { function hdfsFileIsOpenForRead (line 92) | int hdfsFileIsOpenForRead(hdfsFile file) function hdfsGetHedgedReadMetrics (line 97) | int hdfsGetHedgedReadMetrics(hdfsFS fs, struct hdfsHedgedReadMetrics **m... function hdfsFreeHedgedReadMetrics (line 173) | void hdfsFreeHedgedReadMetrics(struct hdfsHedgedReadMetrics *metrics) function hdfsFileGetReadStatistics (line 178) | int hdfsFileGetReadStatistics(hdfsFile file, function hdfsReadStatisticsGetRemoteBytesRead (line 263) | int64_t hdfsReadStatisticsGetRemoteBytesRead( function hdfsFileClearReadStatistics (line 269) | int hdfsFileClearReadStatistics(hdfsFile file) function hdfsFileFreeReadStatistics (line 300) | void hdfsFileFreeReadStatistics(struct hdfsReadStatistics *stats) function hdfsFileIsOpenForWrite (line 305) | int hdfsFileIsOpenForWrite(hdfsFile file) function hdfsFileUsesDirectRead (line 310) | int hdfsFileUsesDirectRead(hdfsFile file) function hdfsFileDisableDirectRead (line 315) | void hdfsFileDisableDirectRead(hdfsFile file) function hdfsDisableDomainSocketSecurity (line 320) | int hdfsDisableDomainSocketSecurity(void) type hdfsJniEnv (line 343) | typedef struct function jthrowable (line 355) | static jthrowable constructNewObjectOfPath(JNIEnv *env, const char *path, function jthrowable (line 376) | static jthrowable hadoopConfGetStr(JNIEnv *env, jobject jConfiguration, function hdfsConfGetStr (line 399) | int hdfsConfGetStr(const char *key, char **val) function hdfsConfStrFree (line 431) | void hdfsConfStrFree(char *val) function jthrowable (line 436) | static jthrowable hadoopConfGetInt(JNIEnv *env, jobject jConfiguration, function hdfsConfGetInt (line 456) | int hdfsConfGetInt(const char *key, int32_t *val) type hdfsBuilderConfOpt (line 488) | struct hdfsBuilderConfOpt { type hdfsBuilder (line 494) | struct hdfsBuilder { type hdfsBuilder (line 503) | struct hdfsBuilder type hdfsBuilder (line 505) | struct hdfsBuilder type hdfsBuilder (line 505) | struct hdfsBuilder function hdfsBuilderConfSetStr (line 513) | int hdfsBuilderConfSetStr(struct hdfsBuilder *bld, const char *key, function hdfsFreeBuilder (line 529) | void hdfsFreeBuilder(struct hdfsBuilder *bld) function hdfsBuilderSetForceNewInstance (line 542) | void hdfsBuilderSetForceNewInstance(struct hdfsBuilder *bld) function hdfsBuilderSetNameNode (line 547) | void hdfsBuilderSetNameNode(struct hdfsBuilder *bld, const char *nn) function hdfsBuilderSetNameNodePort (line 552) | void hdfsBuilderSetNameNodePort(struct hdfsBuilder *bld, tPort port) function hdfsBuilderSetUserName (line 557) | void hdfsBuilderSetUserName(struct hdfsBuilder *bld, const char *userName) function hdfsBuilderSetKerbTicketCachePath (line 562) | void hdfsBuilderSetKerbTicketCachePath(struct hdfsBuilder *bld, function hdfsFS (line 568) | hdfsFS hdfsConnect(const char *host, tPort port) function hdfsFS (line 579) | hdfsFS hdfsConnectNewInstance(const char *host, tPort port) function hdfsFS (line 590) | hdfsFS hdfsConnectAsUser(const char *host, tPort port, const char *user) function hdfsFS (line 602) | hdfsFS hdfsConnectAsUserNewInstance(const char *host, tPort port, function calcEffectiveURI (line 631) | static int calcEffectiveURI(struct hdfsBuilder *bld, char ** uri) type hdfsBuilder (line 671) | struct hdfsBuilder function hdfsFS (line 681) | hdfsFS hdfsBuilderConnect(struct hdfsBuilder *bld) function hdfsDisconnect (line 856) | int hdfsDisconnect(hdfsFS fs) function jthrowable (line 907) | static jthrowable getDefaultBlockSize(JNIEnv *env, jobject jFS, function hdfsFile (line 921) | hdfsFile hdfsOpenFile(hdfsFS fs, const char *path, int flags, type hdfsStreamBuilder (line 937) | struct hdfsStreamBuilder { type hdfsStreamBuilder (line 946) | struct hdfsStreamBuilder type hdfsStreamBuilder (line 950) | struct hdfsStreamBuilder type hdfsStreamBuilder (line 954) | struct hdfsStreamBuilder function hdfsStreamBuilderFree (line 969) | void hdfsStreamBuilderFree(struct hdfsStreamBuilder *bld) function hdfsStreamBuilderSetBufferSize (line 974) | int hdfsStreamBuilderSetBufferSize(struct hdfsStreamBuilder *bld, function hdfsStreamBuilderSetReplication (line 985) | int hdfsStreamBuilderSetReplication(struct hdfsStreamBuilder *bld, function hdfsStreamBuilderSetDefaultBlockSize (line 996) | int hdfsStreamBuilderSetDefaultBlockSize(struct hdfsStreamBuilder *bld, function hdfsFile (line 1007) | static hdfsFile hdfsOpenFileImpl(hdfsFS fs, const char *path, int flags, function hdfsFile (line 1208) | hdfsFile hdfsStreamBuilderBuild(struct hdfsStreamBuilder *bld) function hdfsTruncateFile (line 1218) | int hdfsTruncateFile(hdfsFS fs, const char* path, tOffset newlength) function hdfsUnbufferFile (line 1255) | int hdfsUnbufferFile(hdfsFile file) function hdfsCloseFile (line 1283) | int hdfsCloseFile(hdfsFS fs, hdfsFile file) function hdfsExists (line 1334) | int hdfsExists(hdfsFS fs, const char *path) function readPrepare (line 1375) | static int readPrepare(JNIEnv* env, hdfsFS fs, hdfsFile f, function tSize (line 1396) | tSize hdfsRead(hdfsFS fs, hdfsFile f, void* buffer, tSize length) function tSize (line 1467) | tSize readDirect(hdfsFS fs, hdfsFile f, void* buffer, tSize length) function tSize (line 1508) | tSize hdfsPread(hdfsFS fs, hdfsFile f, tOffset position, function tSize (line 1576) | tSize hdfsWrite(hdfsFS fs, hdfsFile f, const void* buffer, tSize length) function hdfsSeek (line 1647) | int hdfsSeek(hdfsFS fs, hdfsFile f, tOffset desiredPos) function tOffset (line 1682) | tOffset hdfsTell(hdfsFS fs, hdfsFile f) function hdfsFlush (line 1721) | int hdfsFlush(hdfsFS fs, hdfsFile f) function hdfsHFlush (line 1750) | int hdfsHFlush(hdfsFS fs, hdfsFile f) function hdfsHSync (line 1779) | int hdfsHSync(hdfsFS fs, hdfsFile f) function hdfsAvailable (line 1808) | int hdfsAvailable(hdfsFS fs, hdfsFile f) function hdfsCopyImpl (line 1842) | static int hdfsCopyImpl(hdfsFS srcFS, const char *src, hdfsFS dstFS, function hdfsCopy (line 1918) | int hdfsCopy(hdfsFS srcFS, const char *src, hdfsFS dstFS, const char *dst) function hdfsMove (line 1923) | int hdfsMove(hdfsFS srcFS, const char *src, hdfsFS dstFS, const char *dst) function hdfsDelete (line 1928) | int hdfsDelete(hdfsFS fs, const char *path, int recursive) function hdfsRename (line 1973) | int hdfsRename(hdfsFS fs, const char *oldPath, const char *newPath) function hdfsRenameOverwrite (line 2029) | int hdfsRenameOverwrite(hdfsFS fs, const char *oldPath, const char *newP... function hdfsSetWorkingDirectory (line 2186) | int hdfsSetWorkingDirectory(hdfsFS fs, const char *path) function hdfsCreateDirectory (line 2227) | int hdfsCreateDirectory(hdfsFS fs, const char *path) function hdfsSetReplication (line 2278) | int hdfsSetReplication(hdfsFS fs, const char *path, int16_t replication) function hdfsChown (line 2324) | int hdfsChown(hdfsFS fs, const char *path, const char *owner, const char... function hdfsChmod (line 2393) | int hdfsChmod(hdfsFS fs, const char *path, short mode) function hdfsUtime (line 2453) | int hdfsUtime(hdfsFS fs, const char *path, tTime mtime, tTime atime) type hadoopRzOptions (line 2503) | struct hadoopRzOptions type hadoopRzOptions (line 2511) | struct hadoopRzOptions type hadoopRzOptions (line 2513) | struct hadoopRzOptions type hadoopRzOptions (line 2522) | struct hadoopRzOptions function hadoopRzOptionsClearCached (line 2530) | static void hadoopRzOptionsClearCached(JNIEnv *env, function hadoopRzOptionsSetSkipChecksum (line 2540) | int hadoopRzOptionsSetSkipChecksum( function hadoopRzOptionsSetByteBufferPool (line 2554) | int hadoopRzOptionsSetByteBufferPool( function hadoopRzOptionsFree (line 2588) | void hadoopRzOptionsFree(struct hadoopRzOptions *opts) type hadoopRzBuffer (line 2603) | struct hadoopRzBuffer function jthrowable (line 2611) | static jthrowable hadoopRzOptionsGetEnumSet(JNIEnv *env, function hadoopReadZeroExtractBuffer (line 2662) | static int hadoopReadZeroExtractBuffer(JNIEnv *env, function translateZCRException (line 2750) | static int translateZCRException(JNIEnv *env, jthrowable exc) type hadoopRzBuffer (line 2775) | struct hadoopRzBuffer type hadoopRzOptions (line 2776) | struct hadoopRzOptions type hadoopRzBuffer (line 2782) | struct hadoopRzBuffer type hadoopRzBuffer (line 2795) | struct hadoopRzBuffer function hadoopRzBufferLength (line 2848) | int32_t hadoopRzBufferLength(const struct hadoopRzBuffer *buffer) type hadoopRzBuffer (line 2853) | struct hadoopRzBuffer function hadoopRzBufferFree (line 2858) | void hadoopRzBufferFree(hdfsFile file, struct hadoopRzBuffer *buffer) function hdfsFreeHosts (line 3047) | void hdfsFreeHosts(char ***blockHosts) function tOffset (line 3060) | tOffset hdfsGetDefaultBlockSize(hdfsFS fs) function tOffset (line 3088) | tOffset hdfsGetDefaultBlockSizeAtPath(hdfsFS fs, const char *path) function tOffset (line 3122) | tOffset hdfsGetCapacity(hdfsFS fs) function tOffset (line 3162) | tOffset hdfsGetUsed(hdfsFS fs) function getExtendedFileInfoOffset (line 3219) | static size_t getExtendedFileInfoOffset(const char *str) type hdfsExtendedFileInfo (line 3225) | struct hdfsExtendedFileInfo type hdfsExtendedFileInfo (line 3228) | struct hdfsExtendedFileInfo function jthrowable (line 3232) | static jthrowable function jthrowable (line 3382) | static jthrowable function hdfsFileInfo (line 3425) | hdfsFileInfo* hdfsListDirectory(hdfsFS fs, const char *path, int *numEnt... function hdfsFileInfo (line 3521) | hdfsFileInfo *hdfsGetPathInfo(hdfsFS fs, const char *path) function hdfsFreeFileInfoEntry (line 3565) | static void hdfsFreeFileInfoEntry(hdfsFileInfo *hdfsFileInfo) function hdfsFreeFileInfo (line 3573) | void hdfsFreeFileInfo(hdfsFileInfo *hdfsFileInfo, int numEntries) function hdfsFileIsEncrypted (line 3585) | int hdfsFileIsEncrypted(hdfsFileInfo *fileInfo) FILE: native/fs-hdfs/c_src/libhdfs/hdfs.h type hdfsBuilder (line 74) | struct hdfsBuilder type tSize (line 75) | typedef int32_t tSize; type time_t (line 76) | typedef time_t tTime; type tOffset (line 77) | typedef int64_t tOffset; type tPort (line 78) | typedef uint16_t tPort; type tObjectKind (line 79) | typedef enum tObjectKind { type hdfsStreamBuilder (line 83) | struct hdfsStreamBuilder type hdfs_internal (line 89) | struct hdfs_internal type hdfs_internal (line 90) | struct hdfs_internal type hdfsFile_internal (line 92) | struct hdfsFile_internal type hdfsFile_internal (line 93) | struct hdfsFile_internal type hadoopRzOptions (line 95) | struct hadoopRzOptions type hadoopRzBuffer (line 97) | struct hadoopRzBuffer type hdfsReadStatistics (line 117) | struct hdfsReadStatistics { type hdfsReadStatistics (line 139) | struct hdfsReadStatistics type hdfsReadStatistics (line 148) | struct hdfsReadStatistics type hdfsReadStatistics (line 170) | struct hdfsReadStatistics type hdfsHedgedReadMetrics (line 172) | struct hdfsHedgedReadMetrics { type hdfsHedgedReadMetrics (line 191) | struct hdfsHedgedReadMetrics type hdfsHedgedReadMetrics (line 199) | struct hdfsHedgedReadMetrics type hdfsBuilder (line 264) | struct hdfsBuilder type hdfsBuilder (line 281) | struct hdfsBuilder type hdfsBuilder (line 301) | struct hdfsBuilder type hdfsBuilder (line 310) | struct hdfsBuilder type hdfsBuilder (line 319) | struct hdfsBuilder type hdfsBuilder (line 330) | struct hdfsBuilder type hdfsBuilder (line 342) | struct hdfsBuilder type hdfsBuilder (line 356) | struct hdfsBuilder type hdfsStreamBuilder (line 449) | struct hdfsStreamBuilder type hdfsStreamBuilder (line 460) | struct hdfsStreamBuilder type hdfsStreamBuilder (line 476) | struct hdfsStreamBuilder type hdfsStreamBuilder (line 493) | struct hdfsStreamBuilder type hdfsStreamBuilder (line 506) | struct hdfsStreamBuilder type hdfsFileInfo (line 774) | typedef struct { type hadoopRzOptions (line 963) | struct hadoopRzOptions type hadoopRzOptions (line 981) | struct hadoopRzOptions type hadoopRzOptions (line 990) | struct hadoopRzOptions type hadoopRzOptions (line 1016) | struct hadoopRzOptions type hadoopRzBuffer (line 1025) | struct hadoopRzBuffer type hadoopRzBuffer (line 1038) | struct hadoopRzBuffer type hadoopRzBuffer (line 1048) | struct hadoopRzBuffer FILE: native/fs-hdfs/c_src/libhdfs/htable.c type htable_pair (line 27) | struct htable_pair { type htable (line 35) | struct htable { function htable_insert_internal (line 54) | static void htable_insert_internal(struct htable_pair *nelem, function htable_realloc (line 74) | static int htable_realloc(struct htable *htable, uint32_t new_capacity) function round_up_to_power_of_2 (line 97) | static uint32_t round_up_to_power_of_2(uint32_t i) type htable (line 112) | struct htable type htable (line 115) | struct htable function htable_visit (line 135) | void htable_visit(struct htable *htable, visitor_fn_t fun, void *ctx) function htable_free (line 147) | void htable_free(struct htable *htable) function htable_put (line 155) | int htable_put(struct htable *htable, void *key, void *val) function htable_get_internal (line 185) | static int htable_get_internal(const struct htable *htable, type htable (line 213) | struct htable function htable_pop (line 223) | void htable_pop(struct htable *htable, const void *key, function htable_used (line 260) | uint32_t htable_used(const struct htable *htable) function htable_capacity (line 265) | uint32_t htable_capacity(const struct htable *htable) function ht_hash_string (line 270) | uint32_t ht_hash_string(const void *str, uint32_t max) function ht_compare_string (line 282) | int ht_compare_string(const void *a, const void *b) FILE: native/fs-hdfs/c_src/libhdfs/htable.h type htable (line 28) | struct htable type htable (line 59) | struct htable type htable (line 71) | struct htable type htable (line 81) | struct htable type htable (line 96) | struct htable type htable (line 106) | struct htable type htable (line 118) | struct htable type htable (line 128) | struct htable type htable (line 137) | struct htable FILE: native/fs-hdfs/c_src/libhdfs/jni_helper.c type htable (line 32) | struct htable function destroyLocalReference (line 59) | void destroyLocalReference(JNIEnv *env, jobject jObject) function jthrowable (line 65) | static jthrowable validateMethodType(JNIEnv *env, MethType methType) function jthrowable (line 74) | jthrowable newJavaStr(JNIEnv *env, const char *str, jstring *out) function jthrowable (line 94) | jthrowable newCStr(JNIEnv *env, jstring jstr, char **out) function jthrowable (line 111) | jthrowable invokeMethod(JNIEnv *env, jvalue *retval, MethType methType, function jthrowable (line 205) | jthrowable constructNewObjectOfClass(JNIEnv *env, jobject *out, const ch... function jthrowable (line 231) | jthrowable methodIdFromClass(const char *className, const char *methName, function jthrowable (line 260) | jthrowable globalClassReference(const char *className, JNIEnv *env, jcla... function jthrowable (line 308) | jthrowable classNameOfObject(jobject jobj, JNIEnv *env, char **name) function wildcard_expandPath (line 370) | static ssize_t wildcard_expandPath(const char* path, char* expanded) function getClassPath_helper (line 477) | static ssize_t getClassPath_helper(const char *classpath, char* expanded... function JNIEnv (line 642) | static JNIEnv* getGlobalJNIEnv(void) function JNIEnv (line 774) | JNIEnv* getJNIEnv(void) type ThreadLocalState (line 825) | struct ThreadLocalState type ThreadLocalState (line 840) | struct ThreadLocalState function setTLSExceptionStrings (line 853) | void setTLSExceptionStrings(const char *rootCause, const char *stackTrace) function javaObjectIsOfClass (line 872) | int javaObjectIsOfClass(JNIEnv *env, jobject obj, const char *name) function jthrowable (line 888) | jthrowable hadoopConfSetStr(JNIEnv *env, jobject jConfiguration, function jthrowable (line 912) | jthrowable fetchEnumInstance(JNIEnv *env, const char *className, FILE: native/fs-hdfs/c_src/libhdfs/jni_helper.h type MethType (line 58) | typedef enum { FILE: native/fs-hdfs/c_src/libhdfs/os/posix/mutexes.c function init (line 28) | __attribute__((constructor)) static void init() { function mutexLock (line 34) | int mutexLock(mutex *m) { function mutexUnlock (line 43) | int mutexUnlock(mutex *m) { FILE: native/fs-hdfs/c_src/libhdfs/os/posix/platform.h type pthread_mutex_t (line 31) | typedef pthread_mutex_t mutex; type pthread_t (line 32) | typedef pthread_t threadId; FILE: native/fs-hdfs/c_src/libhdfs/os/posix/thread.c function threadCreate (line 37) | int threadCreate(thread *t) { function threadJoin (line 46) | int threadJoin(const thread *t) { FILE: native/fs-hdfs/c_src/libhdfs/os/posix/thread_local_storage.c function hdfsThreadDestructor (line 38) | void hdfsThreadDestructor(void *v) type ThreadLocalState (line 65) | struct ThreadLocalState type ThreadLocalState (line 67) | struct ThreadLocalState type ThreadLocalState (line 68) | struct ThreadLocalState type ThreadLocalState (line 68) | struct ThreadLocalState function threadLocalStorageGet (line 79) | int threadLocalStorageGet(struct ThreadLocalState **state) function threadLocalStorageSet (line 96) | int threadLocalStorageSet(struct ThreadLocalState *state) FILE: native/fs-hdfs/c_src/libhdfs/os/thread.h type thread (line 32) | typedef struct { FILE: native/fs-hdfs/c_src/libhdfs/os/thread_local_storage.h type ThreadLocalState (line 54) | struct ThreadLocalState { type ThreadLocalState (line 76) | struct ThreadLocalState type ThreadLocalState (line 90) | struct ThreadLocalState type ThreadLocalState (line 98) | struct ThreadLocalState FILE: native/fs-hdfs/c_src/libminidfs/native_mini_dfs.c type NativeMiniDfsCluster (line 44) | struct NativeMiniDfsCluster { function hdfsDisableDomainSocketSecurity (line 56) | static int hdfsDisableDomainSocketSecurity(void) function jthrowable (line 75) | static jthrowable nmdConfigureShortCircuit(JNIEnv *env, type NativeMiniDfsCluster (line 106) | struct NativeMiniDfsCluster type NativeMiniDfsConf (line 106) | struct NativeMiniDfsConf type NativeMiniDfsCluster (line 108) | struct NativeMiniDfsCluster type NativeMiniDfsCluster (line 119) | struct NativeMiniDfsCluster function nmdFree (line 215) | void nmdFree(struct NativeMiniDfsCluster* cl) function nmdShutdownInner (line 227) | int nmdShutdownInner(struct NativeMiniDfsCluster* cl, jboolean deleteDfs... function nmdShutdown (line 246) | int nmdShutdown(struct NativeMiniDfsCluster *cl) { function nmdShutdownClean (line 250) | int nmdShutdownClean(struct NativeMiniDfsCluster *cl) { function nmdWaitClusterUp (line 254) | int nmdWaitClusterUp(struct NativeMiniDfsCluster *cl) function nmdGetNameNodePort (line 272) | int nmdGetNameNodePort(const struct NativeMiniDfsCluster *cl) function nmdGetNameNodeHttpAddress (line 294) | int nmdGetNameNodeHttpAddress(const struct NativeMiniDfsCluster *cl, type NativeMiniDfsCluster (line 360) | struct NativeMiniDfsCluster function nmdConfigureHdfsBuilder (line 368) | int nmdConfigureHdfsBuilder(struct NativeMiniDfsCluster *cl, FILE: native/fs-hdfs/c_src/libminidfs/native_mini_dfs.h type hdfsBuilder (line 28) | struct hdfsBuilder type NativeMiniDfsCluster (line 32) | struct NativeMiniDfsCluster type NativeMiniDfsConf (line 39) | struct NativeMiniDfsConf { type NativeMiniDfsCluster (line 68) | struct NativeMiniDfsCluster type NativeMiniDfsConf (line 68) | struct NativeMiniDfsConf type NativeMiniDfsCluster (line 78) | struct NativeMiniDfsCluster type NativeMiniDfsCluster (line 88) | struct NativeMiniDfsCluster type NativeMiniDfsCluster (line 98) | struct NativeMiniDfsCluster type NativeMiniDfsCluster (line 105) | struct NativeMiniDfsCluster type NativeMiniDfsCluster (line 114) | struct NativeMiniDfsCluster type NativeMiniDfsCluster (line 128) | struct NativeMiniDfsCluster type NativeMiniDfsCluster (line 138) | struct NativeMiniDfsCluster type NativeMiniDfsCluster (line 147) | struct NativeMiniDfsCluster type hdfsBuilder (line 148) | struct hdfsBuilder FILE: native/fs-hdfs/src/err.rs type HdfsErr (line 23) | pub enum HdfsErr { method fmt (line 36) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { FILE: native/fs-hdfs/src/hdfs.rs constant O_RDONLY (line 36) | const O_RDONLY: c_int = 0; constant O_WRONLY (line 37) | const O_WRONLY: c_int = 1; constant O_APPEND (line 38) | const O_APPEND: c_int = 8; function get_hdfs_by_full_path (line 46) | pub fn get_hdfs_by_full_path(path: &str) -> Result, HdfsErr> { function get_hdfs (line 51) | pub fn get_hdfs() -> Result, HdfsErr> { function unload_hdfs_cache_by_full_path (line 56) | pub fn unload_hdfs_cache_by_full_path( function unload_hdfs_cache (line 63) | pub fn unload_hdfs_cache(hdfs: Arc) -> Result... type HdfsManager (line 69) | struct HdfsManager { method new (line 74) | fn new() -> Self { method get_hdfs_by_full_path (line 80) | fn get_hdfs_by_full_path(&self, path: &str) -> Result, Hdf... method remove_hdfs_by_full_path (line 123) | fn remove_hdfs_by_full_path( method remove_hdfs (line 135) | fn remove_hdfs(&self, hdfs: Arc) -> Result>... method remove_hdfs_inner (line 139) | fn remove_hdfs_inner(&self, hdfs_key: &str) -> Result &str { method raw (line 170) | pub fn raw(&self) -> hdfsFS { method new_hdfs_file (line 175) | fn new_hdfs_file(&self, path: &str, file: hdfsFile) -> Result Result { method open_with_buf_size (line 197) | pub fn open_with_buf_size( method get_file_status (line 218) | pub fn get_file_status(&self, path: &str) -> Result Result, HdfsE... method default_blocksize (line 258) | pub fn default_blocksize(&self) -> Result { method block_size (line 271) | pub fn block_size(&self, path: &str) -> Result { method capacity (line 287) | pub fn capacity(&self) -> Result { method used (line 298) | pub fn used(&self) -> Result { method exist (line 309) | pub fn exist(&self, path: &str) -> bool { method get_hosts (line 320) | pub fn get_hosts( method create (line 346) | pub fn create(&self, path: &str) -> Result { method create_with_overwrite (line 351) | pub fn create_with_overwrite( method create_with_params (line 359) | pub fn create_with_params( method chmod (line 387) | pub fn chmod(&self, path: &str, mode: i16) -> bool { method chown (line 394) | pub fn chown(&self, path: &str, owner: &str, group: &str) -> bool { method append (line 409) | pub fn append(&self, path: &str) -> Result { method mkdir (line 423) | pub fn mkdir(&self, path: &str) -> Result { method rename (line 438) | pub fn rename( method set_replication (line 467) | pub fn set_replication(&self, path: &str, num: i16) -> Result Result) -> std::fmt::Result { type HdfsFile (line 504) | pub struct HdfsFile { method fs (line 523) | pub fn fs(&self) -> &HdfsFs { method path (line 529) | pub fn path(&self) -> &str { method available (line 533) | pub fn available(&self) -> Result { method close (line 545) | pub fn close(&self) -> Result { method flush (line 557) | pub fn flush(&self) -> bool { method hflush (line 563) | pub fn hflush(&self) -> bool { method hsync (line 570) | pub fn hsync(&self) -> bool { method is_readable (line 575) | pub fn is_readable(&self) -> bool { method is_writable (line 580) | pub fn is_writable(&self) -> bool { method get_file_status (line 585) | pub fn get_file_status(&self) -> Result { method pos (line 590) | pub fn pos(&self) -> Result { method read (line 604) | pub fn read(&self, buf: &mut [u8]) -> Result { method read_with_pos (line 628) | pub fn read_with_pos(&self, pos: i64, buf: &mut [u8]) -> Result bool { method write (line 658) | pub fn write(&self, buf: &[u8]) -> Result { method fmt (line 512) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { type FileStatus (line 689) | pub struct FileStatus { method new (line 698) | fn new(ptr: *const hdfsFileInfo) -> FileStatus { method from_array (line 709) | fn from_array(raw: Arc, idx: u32) -> FileStatus { method ptr (line 719) | fn ptr(&self) -> *const hdfsFileInfo { method name (line 725) | pub fn name(&self) -> &str { method is_file (line 732) | pub fn is_file(&self) -> bool { method is_directory (line 741) | pub fn is_directory(&self) -> bool { method owner (line 750) | pub fn owner(&self) -> &str { method group (line 757) | pub fn group(&self) -> &str { method permission (line 764) | pub fn permission(&self) -> i16 { method len (line 771) | pub fn len(&self) -> usize { method block_size (line 777) | pub fn block_size(&self) -> usize { method replica_count (line 783) | pub fn replica_count(&self) -> i16 { method last_modified (line 789) | pub fn last_modified(&self) -> time_t { method last_access (line 795) | pub fn last_access(&self) -> time_t { type HdfsFileInfoPtr (line 801) | struct HdfsFileInfoPtr { method new (line 814) | fn new(ptr: *const hdfsFileInfo) -> HdfsFileInfoPtr { method new_array (line 818) | pub fn new_array(ptr: *const hdfsFileInfo, len: i32) -> HdfsFileInfoPtr { method drop (line 808) | fn drop(&mut self) { type BlockHosts (line 829) | pub struct BlockHosts { method drop (line 834) | fn drop(&mut self) { constant LOCAL_FS_SCHEME (line 839) | pub const LOCAL_FS_SCHEME: &str = "file"; constant HDFS_FS_SCHEME (line 840) | pub const HDFS_FS_SCHEME: &str = "hdfs"; constant VIEW_FS_SCHEME (line 841) | pub const VIEW_FS_SCHEME: &str = "viewfs"; function get_namenode_uri (line 844) | fn get_namenode_uri(path: &str) -> Result { function get_uri (line 868) | pub fn get_uri(path: &str) -> Result { function test_hdfs_default (line 891) | fn test_hdfs_default() { function test_hdfs (line 908) | fn test_hdfs() { function test_list_status_with_empty_dir (line 981) | fn test_list_status_with_empty_dir() { function test_write_read (line 1021) | fn test_write_read() { FILE: native/fs-hdfs/src/minidfs.rs function get_dfs (line 52) | pub fn get_dfs() -> Arc { type MiniDFS (line 56) | pub struct MiniDFS { method new (line 71) | fn new() -> MiniDFS { method start (line 76) | fn start(conf: &MiniDfsConf) -> Option { method stop (line 83) | fn stop(&self) { method wait_for_clusterup (line 93) | fn wait_for_clusterup(&self) -> bool { method set_hdfs_builder (line 98) | pub fn set_hdfs_builder(&self, builder: *mut hdfsBuilder) -> bool { method namenode_port (line 102) | pub fn namenode_port(&self) -> Option { method namenode_addr (line 109) | pub fn namenode_addr(&self) -> String { method namenode_http_addr (line 117) | pub fn namenode_http_addr(&self) -> Option<(&str, i32)> { method get_hdfs (line 133) | pub fn get_hdfs(&self) -> Result, HdfsErr> { method drop (line 65) | fn drop(&mut self) { function new_mini_dfs_conf (line 138) | fn new_mini_dfs_conf() -> MiniDfsConf { FILE: native/fs-hdfs/src/util.rs type HdfsUtil (line 29) | pub struct HdfsUtil; method copy_file_to_hdfs (line 33) | pub fn copy_file_to_hdfs( method copy_file_from_hdfs (line 47) | pub fn copy_file_from_hdfs( method mv_file_to_hdfs (line 61) | pub fn mv_file_to_hdfs( method mv_file_from_hdfs (line 75) | pub fn mv_file_from_hdfs( method copy (line 95) | pub fn copy( method mv (line 128) | pub fn mv( function test_from_local (line 164) | fn test_from_local() { function test_to_local (line 208) | fn test_to_local() { FILE: native/fs-hdfs/src/walkdir/mod.rs type HdfsWalkDir (line 28) | pub struct HdfsWalkDir { method new (line 35) | pub fn new(root: String) -> Result { method new_with_hdfs (line 40) | pub fn new_with_hdfs(root: String, hdfs: Arc) -> Self { method min_depth (line 56) | pub fn min_depth(mut self, depth: usize) -> Self { method max_depth (line 73) | pub fn max_depth(mut self, depth: usize) -> Self { type Item (line 83) | type Item = Result; type IntoIter (line 84) | type IntoIter = TreeIter; method into_iter (line 86) | fn into_iter(self) -> TreeIter { type HdfsTreeManager (line 97) | struct HdfsTreeManager { method to_value (line 102) | fn to_value(&self, v: String) -> Result { method get_children (line 106) | fn get_children(&self, n: &FileStatus) -> Result, Hdfs... method is_leaf (line 110) | fn is_leaf(&self, n: &FileStatus) -> bool { function test_hdfs_file_list (line 124) | fn test_hdfs_file_list() -> Result<(), HdfsErr> { function set_up_hdfs_env (line 181) | fn set_up_hdfs_env() -> Result, HdfsErr> { FILE: native/fs-hdfs/src/walkdir/tree_iter.rs type TreeIter (line 21) | pub struct TreeIter { function new (line 37) | pub fn new( function next_item (line 51) | fn next_item(&mut self) -> Result, E> { type Item (line 116) | type Item = Result; method next (line 118) | fn next(&mut self) -> Option { type TreeNode (line 123) | pub struct TreeNode { type IterOptions (line 128) | pub struct IterOptions { method fmt (line 134) | fn fmt(&self, f: &mut fmt::Formatter<'_>) -> std::fmt::Result { type TreeManager (line 142) | pub trait TreeManager: Send + Sync { method to_value (line 143) | fn to_value(&self, v: V) -> Result; method get_children (line 145) | fn get_children(&self, n: &N) -> Result, E>; method is_leaf (line 147) | fn is_leaf(&self, n: &N) -> bool; function test_tree_iter (line 158) | fn test_tree_iter() -> Result<(), Error> { function create_test_tree_manager (line 213) | fn create_test_tree_manager() -> TestTreeManager { type TestTreeManager (line 233) | struct TestTreeManager { method to_value (line 238) | fn to_value(&self, v: String) -> Result { method get_children (line 242) | fn get_children(&self, n: &String) -> Result, Error> { method is_leaf (line 255) | fn is_leaf(&self, n: &String) -> bool { FILE: native/hdfs/src/object_store/hdfs.rs type HadoopFileSystem (line 46) | pub struct HadoopFileSystem { method new (line 60) | pub fn new(full_path: &str) -> Option { method path_to_filesystem (line 67) | fn path_to_filesystem(location: &Path) -> String { method get_path_root (line 71) | pub fn get_path_root(&self) -> String { method get_path (line 75) | pub fn get_path(&self, full_path: &str) -> Path { method get_hdfs_host (line 79) | pub fn get_hdfs_host(&self) -> String { method read_range (line 90) | fn read_range(range: &Range, file: &HdfsFile) -> Result { method default (line 51) | fn default() -> Self { method fmt (line 124) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method put_opts (line 131) | async fn put_opts( method put_multipart_opts (line 140) | async fn put_multipart_opts( method get_opts (line 148) | async fn get_opts(&self, location: &Path, options: GetOptions) -> Result... method get_ranges (line 197) | async fn get_ranges(&self, location: &Path, ranges: &[Range]) -> Re... method delete_stream (line 214) | fn delete_stream( method list (line 235) | fn list(&self, prefix: Option<&Path>) -> BoxStream<'static, Result) -> Result Path { function convert_metadata (line 384) | pub fn convert_metadata(file: FileStatus, prefix: &str) -> ObjectMeta { function last_modified (line 394) | fn last_modified(file: &FileStatus) -> DateTime { function check_modified (line 398) | fn check_modified( function convert_walkdir_result (line 424) | fn convert_walkdir_result( constant HDFS_COALESCE_DEFAULT (line 440) | pub const HDFS_COALESCE_DEFAULT: u64 = 1024 * 1024; constant OBJECT_STORE_COALESCE_PARALLEL (line 443) | pub const OBJECT_STORE_COALESCE_PARALLEL: usize = 10; function coalesce_ranges (line 447) | pub async fn coalesce_ranges( function maybe_spawn_blocking (line 479) | pub async fn maybe_spawn_blocking(f: F) -> Result function merge_ranges (line 495) | fn merge_ranges(ranges: &[Range], coalesce: u64) -> Vec> { function to_error (line 532) | fn to_error(err: HdfsErr) -> Error { function test_coalesce_ranges (line 565) | async fn test_coalesce_ranges() { FILE: native/jni-bridge/src/batch_iterator.rs type CometBatchIterator (line 29) | pub struct CometBatchIterator<'a> { constant JVM_CLASS (line 42) | pub const JVM_CLASS: &'static str = "org/apache/comet/CometBatchIterator"; function new (line 44) | pub fn new(env: &mut Env<'a>) -> JniResult> { FILE: native/jni-bridge/src/comet_exec.rs type CometExec (line 27) | pub struct CometExec<'a> { constant JVM_CLASS (line 54) | pub const JVM_CLASS: &'static str = "org/apache/spark/sql/comet/CometSca... function new (line 56) | pub fn new(env: &mut Env<'a>) -> JniResult> { FILE: native/jni-bridge/src/comet_metric_node.rs type CometMetricNode (line 29) | pub struct CometMetricNode<'a> { constant JVM_CLASS (line 40) | pub const JVM_CLASS: &'static str = "org/apache/spark/sql/comet/CometMet... function new (line 42) | pub fn new(env: &mut Env<'a>) -> JniResult> { FILE: native/jni-bridge/src/comet_task_memory_manager.rs type CometTaskMemoryManager (line 30) | pub struct CometTaskMemoryManager<'a> { constant JVM_CLASS (line 40) | pub const JVM_CLASS: &'static str = "org/apache/spark/CometTaskMemoryMan... function new (line 42) | pub fn new(env: &mut Env<'a>) -> JniResult> { FILE: native/jni-bridge/src/errors.rs type ExecutionError (line 53) | pub enum ExecutionError { method from (line 228) | fn from(value: CometError) -> Self { method from (line 258) | fn from(error: prost::DecodeError) -> ExecutionError { method from (line 264) | fn from(error: prost::UnknownEnumValue) -> ExecutionError { method from (line 270) | fn from(error: ArrowError) -> ExecutionError { method from (line 288) | fn from(value: DataFusionError) -> Self { type CometError (line 80) | pub enum CometError { method from (line 187) | fn from(e: Box) -> Self { method from (line 201) | fn from(value: DataFusionError) -> Self { method to_exception (line 306) | fn to_exception(&self) -> Exception { function init (line 174) | pub fn init() { method from (line 210) | fn from(value: CometError) -> Self { method from (line 219) | fn from(value: CometError) -> Self { method from (line 282) | fn from(error: ExpressionError) -> ArrowError { method from (line 294) | fn from(value: ExecutionError) -> Self { method from (line 300) | fn from(value: ExpressionError) -> Self { type ExpressionError (line 346) | pub enum ExpressionError { method from (line 246) | fn from(error: prost::DecodeError) -> ExpressionError { method from (line 252) | fn from(error: prost::UnknownEnumValue) -> ExpressionError { method from (line 276) | fn from(error: ArrowError) -> ExpressionError { type CometResult (line 365) | pub type CometResult = result::Result; type JNIDefault (line 385) | pub trait JNIDefault { method default (line 386) | fn default() -> Self; method default (line 390) | fn default() -> jboolean { method default (line 396) | fn default() -> jbyte { method default (line 402) | fn default() -> jchar { method default (line 408) | fn default() -> jdouble { method default (line 414) | fn default() -> jfloat { method default (line 420) | fn default() -> jint { method default (line 426) | fn default() -> jlong { method default (line 433) | fn default() -> jobject { method default (line 439) | fn default() -> jshort { method default (line 445) | fn default() {} function unwrap_or_throw_default (line 451) | pub fn unwrap_or_throw_default( function throw_exception (line 468) | fn throw_exception(env: &mut Env, error: &CometError, backtrace: Option<... function throw_spark_error_as_json (line 566) | fn throw_spark_error_as_json(env: &mut Env, spark_error: &SparkError) ->... function try_convert_duplicate_field_error (line 583) | fn try_convert_duplicate_field_error(error_msg: &str) -> Option Result... function try_unwrap_or_throw (line 674) | pub fn try_unwrap_or_throw(env: &EnvUnowned, f: F) -> T function jvm (line 713) | pub fn jvm() -> &'static Arc { function error_from_panic (line 759) | pub fn error_from_panic() { function object_result (line 781) | pub fn object_result() { function jlong_result (line 803) | pub fn jlong_result() { function jlong_panic_exception (line 823) | pub fn jlong_panic_exception() { function jlong_result_ok (line 847) | pub fn jlong_result_ok() { function jlong_result_err (line 868) | pub fn jlong_result_err() { function jint_array_result (line 893) | pub fn jint_array_result() { function jint_array_panic_exception (line 918) | pub fn jint_array_panic_exception() { function stacktrace_string (line 946) | pub fn stacktrace_string() { function read_resource (line 962) | fn read_resource(path: &str) -> Result { function Java_Errors_hello (line 976) | pub extern "system" fn Java_Errors_hello( function Java_Errors_div (line 996) | pub extern "system" fn Java_Errors_div( function Java_Errors_div_with_parse (line 1006) | pub extern "system" fn Java_Errors_div_with_parse( function Java_Errors_array_div (line 1023) | pub extern "system" fn Java_Errors_array_div( function assert_pending_java_exception_detailed (line 1043) | fn assert_pending_java_exception_detailed( function assert_exception_type (line 1062) | fn assert_exception_type(env: &mut Env, exception: &JThrowable, expected... function assert_exception_message (line 1085) | fn assert_exception_message(env: &mut Env, exception: JThrowable, expect... FILE: native/jni-bridge/src/lib.rs type StringWrapper (line 129) | pub struct StringWrapper<'a> { function new (line 134) | pub fn new(value: JObject<'a>) -> StringWrapper<'a> { function get (line 138) | pub fn get(&self) -> &JObject<'_> { type BinaryWrapper (line 143) | pub struct BinaryWrapper<'a> { function new (line 148) | pub fn new(value: JObject<'a>) -> BinaryWrapper<'a> { function get (line 152) | pub fn get(&self) -> &JObject<'_> { type Error (line 158) | type Error = Error; function try_from (line 160) | fn try_from(value: JValueOwned<'a>) -> Result, Error> { type Error (line 169) | type Error = Error; function try_from (line 171) | fn try_from(value: JValueOwned<'a>) -> Result, Error> { type JVMClasses (line 193) | pub struct JVMClasses<'a> { function init (line 231) | pub fn init(env: &mut Env) { function get (line 294) | pub fn get() -> &'static JVMClasses<'static> { function with_env (line 303) | pub fn with_env(f: F) -> Result function check_exception (line 324) | pub fn check_exception(env: &mut Env) -> CometResult> { function get_throwable_class_name (line 342) | fn get_throwable_class_name( function get_throwable_message (line 373) | fn get_throwable_message( function convert_exception (line 420) | pub fn convert_exception(env: &mut Env, throwable: &JThrowable) -> Comet... FILE: native/jni-bridge/src/shuffle_block_iterator.rs type CometShuffleBlockIterator (line 29) | pub struct CometShuffleBlockIterator<'a> { constant JVM_CLASS (line 40) | pub const JVM_CLASS: &'static str = "org/apache/comet/CometShuffleBlockI... function new (line 42) | pub fn new(env: &mut Env<'a>) -> JniResult> { FILE: native/proto/build.rs function main (line 22) | fn main() -> Result<()> { FILE: native/shuffle/benches/row_columnar.rs constant BATCH_SIZE (line 32) | const BATCH_SIZE: usize = 5000; constant INT64_SIZE (line 35) | const INT64_SIZE: usize = 8; constant UNSAFE_ROW_POINTER_SIZE (line 40) | const UNSAFE_ROW_POINTER_SIZE: usize = 8; constant ARRAY_HEADER_SIZE (line 43) | const ARRAY_HEADER_SIZE: usize = 8; function write_pointer (line 48) | fn write_pointer(data: &mut [u8], pos: usize, offset: usize, size: usize) { function null_bitset_size (line 54) | fn null_bitset_size(n: usize) -> usize { function make_struct_schema (line 66) | fn make_struct_schema(depth: usize, num_leaf_fields: usize) -> ArrowData... function make_list_schema (line 77) | fn make_list_schema() -> ArrowDataType { function make_map_schema (line 81) | fn make_map_schema() -> ArrowDataType { function build_struct_row (line 97) | fn build_struct_row(depth: usize, num_leaf_fields: usize) -> Vec { function build_list_row (line 153) | fn build_list_row(num_elements: usize) -> Vec { function build_map_row (line 178) | fn build_map_row(num_entries: usize) -> Vec { function run_benchmark (line 221) | fn run_benchmark( function benchmark_primitive_columns (line 274) | fn benchmark_primitive_columns(c: &mut Criterion) { function benchmark_struct_conversion (line 306) | fn benchmark_struct_conversion(c: &mut Criterion) { function benchmark_list_conversion (line 333) | fn benchmark_list_conversion(c: &mut Criterion) { function benchmark_map_conversion (line 358) | fn benchmark_map_conversion(c: &mut Criterion) { function config (line 380) | fn config() -> Criterion { FILE: native/shuffle/benches/shuffle_writer.rs function criterion_benchmark (line 40) | fn criterion_benchmark(c: &mut Criterion) { function create_shuffle_writer_exec (line 139) | fn create_shuffle_writer_exec( function create_batches (line 160) | fn create_batches(size: usize, count: usize) -> Vec { function create_batch (line 169) | fn create_batch(num_rows: usize, allow_nulls: bool) -> RecordBatch { function config (line 203) | fn config() -> Criterion { FILE: native/shuffle/src/bin/shuffle_bench.rs type Args (line 60) | struct Args { function main (line 119) | fn main() { function print_shuffle_metrics (line 255) | fn print_shuffle_metrics(metrics: &MetricsSet, total_wall_time_secs: f64) { function print_input_metrics (line 304) | fn print_input_metrics(metrics: &MetricsSet) { function read_parquet_metadata (line 336) | fn read_parquet_metadata(path: &Path, limit: usize) -> (SchemaRef, u64) { function collect_parquet_paths (line 364) | fn collect_parquet_paths(path: &Path) -> Vec { function run_shuffle_write (line 387) | fn run_shuffle_write( function execute_shuffle_write (line 429) | async fn execute_shuffle_write( function collect_input_metrics (line 494) | fn collect_input_metrics(exec: &ShuffleWriterExec) -> MetricsSet { function run_concurrent_shuffle_writes (line 514) | fn run_concurrent_shuffle_writes( function build_partitioning (line 575) | fn build_partitioning( function parse_codec (line 602) | fn parse_codec(codec: &str, zstd_level: i32) -> CompressionCodec { function parse_hash_columns (line 615) | fn parse_hash_columns(s: &str) -> Vec { function describe_schema (line 622) | fn describe_schema(schema: &arrow::datatypes::Schema) -> String { function format_number (line 653) | fn format_number(n: usize) -> String { function format_bytes (line 665) | fn format_bytes(bytes: usize) -> String { FILE: native/shuffle/src/comet_partitioning.rs type CometPartitioning (line 24) | pub enum CometPartitioning { method partition_count (line 42) | pub fn partition_count(&self) -> usize { function pmod (line 51) | pub(crate) fn pmod(hash: u32, n: usize) -> usize { function test_pmod (line 64) | fn test_pmod() { FILE: native/shuffle/src/ipc.rs function read_ipc_compressed (line 23) | pub fn read_ipc_compressed(bytes: &[u8]) -> Result { FILE: native/shuffle/src/metrics.rs type ShufflePartitionerMetrics (line 23) | pub(crate) struct ShufflePartitionerMetrics { method new (line 50) | pub(crate) fn new(metrics: &ExecutionPlanMetricsSet, partition: usize)... FILE: native/shuffle/src/partitioners/empty_schema.rs type EmptySchemaShufflePartitioner (line 32) | pub(crate) struct EmptySchemaShufflePartitioner { method try_new (line 43) | pub(crate) fn try_new( method insert_batch (line 70) | async fn insert_batch(&mut self, batch: RecordBatch) -> datafusion::comm... method shuffle_write (line 85) | fn shuffle_write(&mut self) -> datafusion::common::Result<()> { FILE: native/shuffle/src/partitioners/multi_partition.rs type ScratchSpace (line 44) | struct ScratchSpace { method map_partition_ids_to_starts_and_indices (line 60) | fn map_partition_ids_to_starts_and_indices( type MultiPartitionShuffleRepartitioner (line 108) | pub(crate) struct MultiPartitionShuffleRepartitioner { method try_new (line 132) | pub(crate) fn try_new( method partitioning_batch (line 200) | async fn partitioning_batch(&mut self, input: RecordBatch) -> datafusi... method buffer_partitioned_batch_may_spill (line 395) | async fn buffer_partitioned_batch_may_spill( method shuffle_write_partition (line 437) | fn shuffle_write_partition( method used (line 460) | fn used(&self) -> usize { method spilled_bytes (line 464) | fn spilled_bytes(&self) -> usize { method spill_count (line 468) | fn spill_count(&self) -> usize { method data_size (line 472) | fn data_size(&self) -> usize { method partitioned_batches (line 479) | fn partitioned_batches(&mut self) -> PartitionedBatchesProducer { method spill (line 490) | pub(crate) fn spill(&mut self) -> datafusion::common::Result<()> { method partition_writers (line 527) | pub(crate) fn partition_writers(&self) -> &[PartitionWriter] { method insert_batch (line 537) | async fn insert_batch(&mut self, batch: RecordBatch) -> datafusion::comm... method shuffle_write (line 558) | fn shuffle_write(&mut self) -> datafusion::common::Result<()> { method fmt (line 635) | fn fmt(&self, f: &mut Formatter<'_>) -> fmt::Result { FILE: native/shuffle/src/partitioners/partitioned_batch_iterator.rs type PartitionedBatchesProducer (line 25) | pub(super) struct PartitionedBatchesProducer { method new (line 32) | pub(super) fn new( method produce (line 44) | pub(super) fn produce(&mut self, partition_id: usize) -> PartitionedBa... type PartitionedBatchIterator (line 54) | pub(crate) struct PartitionedBatchIterator<'a> { function new (line 62) | fn new( type Item (line 91) | type Item = datafusion::common::Result; method next (line 93) | fn next(&mut self) -> Option { FILE: native/shuffle/src/partitioners/single_partition.rs type SinglePartitionShufflePartitioner (line 30) | pub(crate) struct SinglePartitionShufflePartitioner { method try_new (line 45) | pub(crate) fn try_new( method add_buffered_batch (line 81) | fn add_buffered_batch(&mut self, batch: RecordBatch) { method concat_buffered_batches (line 87) | fn concat_buffered_batches(&mut self) -> datafusion::common::Result datafusion::comm... method shuffle_write (line 157) | fn shuffle_write(&mut self) -> datafusion::common::Result<()> { FILE: native/shuffle/src/partitioners/traits.rs type ShufflePartitioner (line 22) | pub(crate) trait ShufflePartitioner: Send + Sync { method insert_batch (line 24) | async fn insert_batch(&mut self, batch: RecordBatch) -> Result<()>; method shuffle_write (line 26) | fn shuffle_write(&mut self) -> Result<()>; FILE: native/shuffle/src/shuffle_writer.rs type ShuffleWriterExec (line 52) | pub struct ShuffleWriterExec { method try_new (line 75) | pub fn try_new( method fmt_as (line 106) | fn fmt_as(&self, t: DisplayFormatType, f: &mut Formatter) -> fmt::Result { method as_any (line 123) | fn as_any(&self) -> &dyn Any { method name (line 127) | fn name(&self) -> &str { method metrics (line 131) | fn metrics(&self) -> Option { method properties (line 135) | fn properties(&self) -> &Arc { method schema (line 140) | fn schema(&self) -> SchemaRef { method children (line 144) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 148) | fn with_new_children( method execute (line 166) | fn execute( function external_shuffle (line 197) | async fn external_shuffle( function roundtrip_ipc (line 289) | fn roundtrip_ipc() { function test_single_partition_shuffle_writer (line 314) | fn test_single_partition_shuffle_writer() { function test_insert_larger_batch (line 321) | fn test_insert_larger_batch() { function test_insert_smaller_batch (line 327) | fn test_insert_smaller_batch() { function test_large_number_of_partitions (line 334) | fn test_large_number_of_partitions() { function test_large_number_of_partitions_spilling (line 341) | fn test_large_number_of_partitions_spilling() { function shuffle_partitioner_memory (line 346) | async fn shuffle_partitioner_memory() { function create_runtime (line 392) | fn create_runtime(memory_limit: usize) -> Arc { function shuffle_write_test (line 401) | fn shuffle_write_test( function create_batch (line 489) | fn create_batch(batch_size: usize) -> RecordBatch { function test_round_robin_deterministic (line 501) | fn test_round_robin_deterministic() { function test_batch_coalescing_reduces_size (line 594) | fn test_batch_coalescing_reduces_size() { function read_all_ipc_blocks (line 678) | fn read_all_ipc_blocks(data: &[u8]) -> usize { function test_empty_schema_shuffle_writer (line 700) | fn test_empty_schema_shuffle_writer() { function test_empty_schema_shuffle_writer_zero_rows (line 790) | fn test_empty_schema_shuffle_writer_zero_rows() { FILE: native/shuffle/src/spark_crc32c_hasher.rs type SparkCrc32cHasher (line 31) | pub struct SparkCrc32cHasher { method new (line 39) | pub fn new(initial: u32) -> Self { method finalize (line 43) | pub fn finalize(&self) -> u32 { method finish (line 49) | fn finish(&self) -> u64 { method write (line 53) | fn write(&mut self, bytes: &[u8]) { constant TEST_STRING (line 62) | const TEST_STRING: &[u8] = constant CHECKSUM (line 64) | const CHECKSUM: u32 = 0x20_CB_1E_59; function can_hash (line 67) | fn can_hash() { function can_hash_in_chunks (line 76) | fn can_hash_in_chunks() { FILE: native/shuffle/src/spark_unsafe/list.rs type SparkUnsafeArray (line 88) | pub struct SparkUnsafeArray { method new (line 112) | pub fn new(addr: i64) -> Self { method get_num_elements (line 134) | pub(crate) fn get_num_elements(&self) -> usize { method get_header_portion_in_bytes (line 140) | const fn get_header_portion_in_bytes(num_fields: i64) -> i64 { method is_null_at (line 146) | pub(crate) fn is_null_at(&self, index: usize) -> bool { method null_bitset_ptr (line 165) | fn null_bitset_ptr(&self) -> *const i64 { method is_null_in_bitset (line 174) | unsafe fn is_null_in_bitset(null_words: *const i64, idx: usize) -> bool { method append_booleans_to_builder (line 189) | pub(crate) fn append_booleans_to_builder( method append_timestamps_to_builder (line 233) | pub(crate) fn append_timestamps_to_builder( method append_dates_to_builder (line 287) | pub(crate) fn append_dates_to_builder( method get_row_addr (line 96) | fn get_row_addr(&self) -> i64 { method get_element_offset (line 101) | fn get_element_offset(&self, index: usize, element_size: usize) -> *cons... function append_to_builder (line 341) | pub fn append_to_builder( function append_list_element (line 475) | pub fn append_list_element( FILE: native/shuffle/src/spark_unsafe/map.rs type SparkUnsafeMap (line 24) | pub struct SparkUnsafeMap { method new (line 31) | pub(crate) fn new(addr: i64, size: i32) -> Self { function append_map_elements (line 70) | pub fn append_map_elements( function get_map_key_value_fields (line 90) | pub fn get_map_key_value_fields( FILE: native/shuffle/src/spark_unsafe/row.rs constant NESTED_TYPE_BUILDER_CAPACITY (line 49) | const NESTED_TYPE_BUILDER_CAPACITY: usize = 100; type SparkUnsafeRow (line 52) | pub struct SparkUnsafeRow { method new (line 91) | fn new(schema: &[DataType]) -> Self { method is_null_row (line 100) | pub fn is_null_row(&self) -> bool { method get_row_bitset_width (line 107) | pub const fn get_row_bitset_width(num_fields: usize) -> usize { method new_with_num_fields (line 111) | pub fn new_with_num_fields(num_fields: usize) -> Self { method point_to_slice (line 120) | pub fn point_to_slice(&mut self, slice: &[u8]) { method point_to (line 126) | pub(crate) fn point_to(&mut self, row_addr: i64, row_size: i32) { method get_row_size (line 131) | pub fn get_row_size(&self) -> i32 { method is_null_at (line 137) | pub(crate) fn is_null_at(&self, index: usize) -> bool { method set_not_null_at (line 153) | pub fn set_not_null_at(&mut self, index: usize) { method get_row_addr (line 59) | fn get_row_addr(&self) -> i64 { method get_element_offset (line 63) | fn get_element_offset(&self, index: usize, element_size: usize) -> *cons... method default (line 81) | fn default() -> Self { function append_field (line 209) | pub(super) fn append_field( function append_nested_struct_fields_field_major (line 362) | fn append_nested_struct_fields_field_major( function append_list_column_batch (line 591) | fn append_list_column_batch( function append_map_column_batch (line 675) | fn append_map_column_batch( function append_struct_fields_field_major (line 784) | fn append_struct_fields_field_major( function append_columns (line 1018) | fn append_columns( function make_builders (line 1216) | fn make_builders( function process_sorted_row_partition (line 1297) | pub fn process_sorted_row_partition( function builder_to_array (line 1390) | fn builder_to_array( function make_batch (line 1440) | fn make_batch(arrays: Vec, row_count: usize) -> Result i64; method get_element_offset (line 56) | fn get_element_offset(&self, index: usize, element_size: usize) -> *co... method get_boolean (line 58) | fn get_boolean(&self, index: usize) -> bool; method get_byte (line 59) | fn get_byte(&self, index: usize) -> i8; method get_short (line 60) | fn get_short(&self, index: usize) -> i16; method get_int (line 61) | fn get_int(&self, index: usize) -> i32; method get_long (line 62) | fn get_long(&self, index: usize) -> i64; method get_float (line 63) | fn get_float(&self, index: usize) -> f32; method get_double (line 64) | fn get_double(&self, index: usize) -> f64; method get_date (line 65) | fn get_date(&self, index: usize) -> i32; method get_timestamp (line 66) | fn get_timestamp(&self, index: usize) -> i64; method get_offset_and_len (line 70) | fn get_offset_and_len(&self, index: usize) -> (i32, i32) { method get_string (line 78) | fn get_string(&self, index: usize) -> &str { method get_binary (line 94) | fn get_binary(&self, index: usize) -> &[u8] { method get_decimal (line 108) | fn get_decimal(&self, index: usize, precision: u8) -> i128 { method get_struct (line 118) | fn get_struct(&self, index: usize, num_fields: usize) -> SparkUnsafeRow { method get_array (line 127) | fn get_array(&self, index: usize) -> SparkUnsafeArray { method get_map (line 132) | fn get_map(&self, index: usize) -> SparkUnsafeMap { FILE: native/shuffle/src/writers/buf_batch_writer.rs type BufBatchWriter (line 32) | pub(crate) struct BufBatchWriter, W: Write> { function new (line 45) | pub(crate) fn new( function write (line 61) | pub(crate) fn write( function write_batch_to_buffer (line 87) | fn write_batch_to_buffer( function flush (line 109) | pub(crate) fn flush( function writer_stream_position (line 139) | pub(crate) fn writer_stream_position(&mut self) -> datafusion::common::R... FILE: native/shuffle/src/writers/checksum.rs type Checksum (line 28) | pub(crate) enum Checksum { method try_new (line 38) | pub(crate) fn try_new(algo: i32, initial_opt: Option) -> CometRes... method update (line 72) | pub(crate) fn update(&mut self, cursor: &mut Cursor<&mut Vec>) -> ... method finalize (line 92) | pub(crate) fn finalize(self) -> u32 { function test_crc32 (line 107) | fn test_crc32() { function test_adler32 (line 122) | fn test_adler32() { function test_crc32c (line 137) | fn test_crc32c() { FILE: native/shuffle/src/writers/shuffle_block_writer.rs type CompressionCodec (line 28) | pub enum CompressionCodec { type ShuffleBlockWriter (line 37) | pub struct ShuffleBlockWriter { method try_new (line 43) | pub fn try_new(schema: &Schema, codec: CompressionCodec) -> Result( FILE: native/shuffle/src/writers/spill.rs type SpillFile (line 28) | struct SpillFile { type PartitionWriter (line 34) | pub(crate) struct PartitionWriter { method try_new (line 44) | pub(crate) fn try_new( method ensure_spill_file_created (line 53) | fn ensure_spill_file_created( method spill (line 78) | pub(crate) fn spill( method path (line 116) | pub(crate) fn path(&self) -> Option<&std::path::Path> { method has_spill_file (line 123) | pub(crate) fn has_spill_file(&self) -> bool { FILE: native/spark-expr/benches/aggregate.rs function criterion_benchmark (line 42) | fn criterion_benchmark(c: &mut Criterion) { function agg_test (line 272) | async fn agg_test( function create_aggregate (line 292) | fn create_aggregate( function create_record_batch (line 321) | fn create_record_batch(num_rows: usize) -> RecordBatch { function create_int64_record_batch (line 346) | fn create_int64_record_batch(num_rows: usize) -> RecordBatch { function config (line 371) | fn config() -> Criterion { FILE: native/spark-expr/benches/bloom_filter_agg.rs function criterion_benchmark (line 40) | fn criterion_benchmark(c: &mut Criterion) { function agg_test (line 85) | async fn agg_test( function create_aggregate (line 105) | fn create_aggregate( function create_record_batch (line 134) | fn create_record_batch(num_rows: usize) -> RecordBatch { function config (line 152) | fn config() -> Criterion { FILE: native/spark-expr/benches/cast_from_boolean.rs function criterion_benchmark (line 26) | fn criterion_benchmark(c: &mut Criterion) { function create_boolean_batch (line 114) | fn create_boolean_batch() -> RecordBatch { function config (line 128) | fn config() -> Criterion { FILE: native/spark-expr/benches/cast_from_string.rs function criterion_benchmark (line 25) | fn criterion_benchmark(c: &mut Criterion) { function create_small_int_string_batch (line 137) | fn create_small_int_string_batch() -> RecordBatch { function create_int_string_batch (line 152) | fn create_int_string_batch() -> RecordBatch { function create_decimal_string_batch (line 167) | fn create_decimal_string_batch() -> RecordBatch { function create_decimal_cast_string_batch (line 185) | fn create_decimal_cast_string_batch() -> RecordBatch { function config (line 229) | fn config() -> Criterion { FILE: native/spark-expr/benches/cast_int_to_timestamp.rs constant BATCH_SIZE (line 26) | const BATCH_SIZE: usize = 8192; function criterion_benchmark (line 28) | fn criterion_benchmark(c: &mut Criterion) { function create_int8_batch (line 94) | fn create_int8_batch() -> RecordBatch { function create_int16_batch (line 107) | fn create_int16_batch() -> RecordBatch { function create_int32_batch (line 120) | fn create_int32_batch() -> RecordBatch { function create_int64_batch (line 133) | fn create_int64_batch() -> RecordBatch { function config (line 146) | fn config() -> Criterion { FILE: native/spark-expr/benches/cast_non_int_numeric_timestamp.rs constant BATCH_SIZE (line 26) | const BATCH_SIZE: usize = 8192; function criterion_benchmark (line 28) | fn criterion_benchmark(c: &mut Criterion) { function create_float32_batch (line 93) | fn create_float32_batch() -> RecordBatch { function create_float64_batch (line 106) | fn create_float64_batch() -> RecordBatch { function create_boolean_batch (line 119) | fn create_boolean_batch() -> RecordBatch { function create_decimal128_batch (line 132) | fn create_decimal128_batch() -> RecordBatch { function config (line 150) | fn config() -> Criterion { FILE: native/spark-expr/benches/cast_numeric.rs function criterion_benchmark (line 25) | fn criterion_benchmark(c: &mut Criterion) { function create_int32_batch (line 57) | fn create_int32_batch() -> RecordBatch { function config (line 72) | fn config() -> Criterion { FILE: native/spark-expr/benches/conditional.rs function make_col (line 33) | fn make_col(name: &str, index: usize) -> Arc { function make_lit_i32 (line 37) | fn make_lit_i32(n: i32) -> Arc { function make_null_lit (line 41) | fn make_null_lit() -> Arc { function criterion_benchmark (line 45) | fn criterion_benchmark(c: &mut Criterion) { FILE: native/spark-expr/benches/date_trunc.rs function criterion_benchmark (line 23) | fn criterion_benchmark(c: &mut Criterion) { function create_date_array (line 39) | fn create_date_array() -> Date32Array { function config (line 46) | fn config() -> Criterion { FILE: native/spark-expr/benches/decimal_div.rs function criterion_benchmark (line 27) | fn criterion_benchmark(c: &mut Criterion) { FILE: native/spark-expr/benches/normalize_nan.rs constant BATCH_SIZE (line 30) | const BATCH_SIZE: usize = 8192; function make_col (line 32) | fn make_col(name: &str, index: usize) -> Arc { function create_float_batch (line 37) | fn create_float_batch(nan_pct: usize, neg_zero_pct: usize, null_pct: usi... function bench_normalize_nan_and_zero (line 58) | fn bench_normalize_nan_and_zero(c: &mut Criterion) { FILE: native/spark-expr/benches/padding.rs function create_string_array (line 27) | fn create_string_array(size: usize) -> ArrayRef { function criterion_benchmark (line 39) | fn criterion_benchmark(c: &mut Criterion) { FILE: native/spark-expr/benches/to_csv.rs function create_struct_array (line 27) | fn create_struct_array(array_size: usize) -> StructArray { function criterion_benchmark (line 81) | fn criterion_benchmark(c: &mut Criterion) { FILE: native/spark-expr/benches/wide_decimal.rs constant BATCH_SIZE (line 33) | const BATCH_SIZE: usize = 8192; function make_decimal_batch (line 36) | fn make_decimal_batch(p1: u8, s1: i8, p2: u8, s2: i8) -> RecordBatch { function build_old_expr (line 53) | fn build_old_expr( function build_new_expr (line 83) | fn build_new_expr(op: WideDecimalOp, p_out: u8, s_out: i8) -> Arc Result { type Avg (line 47) | pub struct Avg { method new (line 57) | pub fn new(name: impl Into, data_type: DataType) -> Self { method as_any (line 71) | fn as_any(&self) -> &dyn Any { method accumulator (line 75) | fn accumulator(&self, _acc_args: AccumulatorArgs) -> Result Result> { method name (line 102) | fn name(&self) -> &str { method reverse_expr (line 106) | fn reverse_expr(&self) -> ReversedUDAF { method groups_accumulator_supported (line 110) | fn groups_accumulator_supported(&self, _args: AccumulatorArgs) -> bool { method create_groups_accumulator (line 114) | fn create_groups_accumulator( method default_value (line 132) | fn default_value(&self, _data_type: &DataType) -> Result { method signature (line 136) | fn signature(&self) -> &Signature { method return_type (line 140) | fn return_type(&self, arg_types: &[DataType]) -> Result { type AvgAccumulator (line 147) | pub struct AvgAccumulator { method state (line 153) | fn state(&mut self) -> Result> { method update_batch (line 160) | fn update_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method merge_batch (line 170) | fn merge_batch(&mut self, states: &[ArrayRef]) -> Result<()> { method evaluate (line 182) | fn evaluate(&mut self) -> Result { method size (line 194) | fn size(&self) -> usize { type AvgGroupsAccumulator (line 206) | struct AvgGroupsAccumulator function new (line 229) | pub fn new(return_data_type: &DataType, avg_fn: F) -> Self { method update_batch (line 244) | fn update_batch( method merge_batch (line 287) | fn merge_batch( method evaluate (line 316) | fn evaluate(&mut self, emit_to: EmitTo) -> Result { method state (line 334) | fn state(&mut self, emit_to: EmitTo) -> Result> { method size (line 348) | fn size(&self) -> usize { FILE: native/spark-expr/src/agg_funcs/avg_decimal.rs function avg_return_type (line 45) | fn avg_return_type(_name: &str, data_type: &DataType) -> Result { type AvgDecimal (line 60) | pub struct AvgDecimal { method hash (line 82) | fn hash(&self, state: &mut H) { method new (line 92) | pub fn new( method eq (line 71) | fn eq(&self, other: &Self) -> bool { method as_any (line 112) | fn as_any(&self) -> &dyn Any { method accumulator (line 116) | fn accumulator(&self, _acc_args: AccumulatorArgs) -> Result Result> { method name (line 152) | fn name(&self) -> &str { method reverse_expr (line 156) | fn reverse_expr(&self) -> ReversedUDAF { method groups_accumulator_supported (line 160) | fn groups_accumulator_supported(&self, _args: AccumulatorArgs) -> bool { method create_groups_accumulator (line 164) | fn create_groups_accumulator( method default_value (line 191) | fn default_value(&self, _data_type: &DataType) -> Result { method signature (line 203) | fn signature(&self) -> &Signature { method return_type (line 207) | fn return_type(&self, arg_types: &[DataType]) -> Result { method is_nullable (line 211) | fn is_nullable(&self) -> bool { type AvgDecimalAccumulator (line 220) | struct AvgDecimalAccumulator { method new (line 235) | pub fn new( method wrap_error_with_context (line 260) | fn wrap_error_with_context( method update_single (line 273) | fn update_single(&mut self, values: &Decimal128Array, idx: usize) -> R... function make_decimal128 (line 302) | fn make_decimal128(value: Option, precision: u8, scale: i8) -> Sca... method state (line 307) | fn state(&mut self) -> Result> { method update_batch (line 314) | fn update_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method merge_batch (line 338) | fn merge_batch(&mut self, states: &[ArrayRef]) -> Result<()> { method evaluate (line 366) | fn evaluate(&mut self) -> Result { method size (line 402) | fn size(&self) -> usize { type AvgDecimalGroupsAccumulator (line 408) | struct AvgDecimalGroupsAccumulator { method new (line 439) | pub fn new( method wrap_error_with_context (line 467) | fn wrap_error_with_context( method update_single (line 481) | fn update_single(&mut self, group_index: usize, value: i128) -> Result... function ensure_bit_capacity (line 495) | fn ensure_bit_capacity(builder: &mut BooleanBufferBuilder, capacity: usi... method update_batch (line 503) | fn update_batch( method merge_batch (line 540) | fn merge_batch( method evaluate (line 595) | fn evaluate(&mut self, emit_to: EmitTo) -> Result { method state (line 636) | fn state(&mut self, emit_to: EmitTo) -> Result> { method size (line 653) | fn size(&self) -> usize { function avg (line 668) | fn avg(sum: i128, count: i128, target_min: i128, target_max: i128, scale... FILE: native/spark-expr/src/agg_funcs/correlation.rs type Correlation (line 42) | pub struct Correlation { method new (line 49) | pub fn new(name: impl Into, data_type: DataType, null_on_divid... method as_any (line 62) | fn as_any(&self) -> &dyn Any { method name (line 66) | fn name(&self) -> &str { method signature (line 70) | fn signature(&self) -> &Signature { method return_type (line 74) | fn return_type(&self, _arg_types: &[DataType]) -> Result { method default_value (line 77) | fn default_value(&self, _data_type: &DataType) -> Result { method accumulator (line 81) | fn accumulator(&self, _acc_args: AccumulatorArgs) -> Result Result> { type CorrelationAccumulator (line 125) | pub struct CorrelationAccumulator { method try_new (line 134) | pub fn try_new(null_on_divide_by_zero: bool) -> Result { method state (line 145) | fn state(&mut self) -> Result> { method update_batch (line 156) | fn update_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method retract_batch (line 176) | fn retract_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method merge_batch (line 193) | fn merge_batch(&mut self, states: &[ArrayRef]) -> Result<()> { method evaluate (line 219) | fn evaluate(&mut self) -> Result { method size (line 243) | fn size(&self) -> usize { FILE: native/spark-expr/src/agg_funcs/covariance.rs type Covariance (line 40) | pub struct Covariance { method hash (line 48) | fn hash(&self, state: &mut H) { method new (line 58) | pub fn new( method as_any (line 77) | fn as_any(&self) -> &dyn Any { method name (line 81) | fn name(&self) -> &str { method signature (line 85) | fn signature(&self) -> &Signature { method return_type (line 89) | fn return_type(&self, _arg_types: &[DataType]) -> Result { method default_value (line 92) | fn default_value(&self, _data_type: &DataType) -> Result { method accumulator (line 96) | fn accumulator(&self, _acc_args: AccumulatorArgs) -> Result Result> { type CovarianceAccumulator (line 131) | pub struct CovarianceAccumulator { method try_new (line 142) | pub fn try_new(s_type: StatsType, null_on_divide_by_zero: bool) -> Res... method get_count (line 153) | pub fn get_count(&self) -> f64 { method get_mean1 (line 157) | pub fn get_mean1(&self) -> f64 { method get_mean2 (line 161) | pub fn get_mean2(&self) -> f64 { method get_algo_const (line 165) | pub fn get_algo_const(&self) -> f64 { method state (line 171) | fn state(&mut self) -> Result> { method update_batch (line 180) | fn update_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method retract_batch (line 221) | fn retract_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method merge_batch (line 262) | fn merge_batch(&mut self, states: &[ArrayRef]) -> Result<()> { method evaluate (line 289) | fn evaluate(&mut self) -> Result { method size (line 310) | fn size(&self) -> usize { FILE: native/spark-expr/src/agg_funcs/stddev.rs type Stddev (line 40) | pub struct Stddev { method hash (line 48) | fn hash(&self, state: &mut H) { method new (line 58) | pub fn new( method as_any (line 82) | fn as_any(&self) -> &dyn Any { method name (line 86) | fn name(&self) -> &str { method signature (line 90) | fn signature(&self) -> &Signature { method return_type (line 94) | fn return_type(&self, _arg_types: &[DataType]) -> Result { method accumulator (line 98) | fn accumulator(&self, _acc_args: AccumulatorArgs) -> Result Result> { method default_value (line 135) | fn default_value(&self, _data_type: &DataType) -> Result { type StddevAccumulator (line 142) | pub struct StddevAccumulator { method try_new (line 148) | pub fn try_new(s_type: StatsType, null_on_divide_by_zero: bool) -> Res... method get_m2 (line 154) | pub fn get_m2(&self) -> f64 { method state (line 160) | fn state(&mut self) -> Result> { method update_batch (line 168) | fn update_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method retract_batch (line 172) | fn retract_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method merge_batch (line 176) | fn merge_batch(&mut self, states: &[ArrayRef]) -> Result<()> { method evaluate (line 180) | fn evaluate(&mut self) -> Result { method size (line 189) | fn size(&self) -> usize { FILE: native/spark-expr/src/agg_funcs/sum_decimal.rs type SumDecimal (line 33) | pub struct SumDecimal { method hash (line 65) | fn hash(&self, state: &mut H) { method try_new (line 75) | pub fn try_new( method eq (line 53) | fn eq(&self, other: &Self) -> bool { method as_any (line 102) | fn as_any(&self) -> &dyn Any { method accumulator (line 106) | fn accumulator(&self, _args: AccumulatorArgs) -> DFResult DFResult> { method name (line 126) | fn name(&self) -> &str { method signature (line 130) | fn signature(&self) -> &Signature { method return_type (line 134) | fn return_type(&self, _arg_types: &[DataType]) -> DFResult { method groups_accumulator_supported (line 138) | fn groups_accumulator_supported(&self, _args: AccumulatorArgs) -> bool { method create_groups_accumulator (line 142) | fn create_groups_accumulator( method default_value (line 155) | fn default_value(&self, _data_type: &DataType) -> DFResult { method reverse_expr (line 162) | fn reverse_expr(&self) -> ReversedUDAF { method is_nullable (line 166) | fn is_nullable(&self) -> bool { type SumDecimalAccumulator (line 174) | struct SumDecimalAccumulator { method new (line 185) | fn new( method update_single (line 205) | fn update_single(&mut self, values: &Decimal128Array, idx: usize) -> D... method wrap_error_with_context (line 231) | fn wrap_error_with_context(&self, error: crate::SparkError) -> DataFus... method update_batch (line 243) | fn update_batch(&mut self, values: &[ArrayRef]) -> DFResult<()> { method evaluate (line 275) | fn evaluate(&mut self) -> DFResult { method size (line 294) | fn size(&self) -> usize { method state (line 298) | fn state(&mut self) -> DFResult> { method merge_batch (line 313) | fn merge_batch(&mut self, states: &[ArrayRef]) -> DFResult<()> { type SumDecimalGroupsAccumulator (line 371) | struct SumDecimalGroupsAccumulator { method new (line 382) | fn new( method resize_helper (line 400) | fn resize_helper(&mut self, total_num_groups: usize) { method wrap_error_with_context (line 407) | fn wrap_error_with_context(&self, error: crate::SparkError) -> DataFus... method update_single (line 418) | fn update_single(&mut self, group_index: usize, value: i128) -> DFResu... method update_batch (line 442) | fn update_batch( method evaluate (line 477) | fn evaluate(&mut self, emit_to: EmitTo) -> DFResult { method state (line 526) | fn state(&mut self, emit_to: EmitTo) -> DFResult> { method merge_batch (line 540) | fn merge_batch( method size (line 612) | fn size(&self) -> usize { function invalid_data_type (line 637) | fn invalid_data_type() { function sum_no_overflow (line 648) | async fn sum_no_overflow() -> Result<()> { function create_record_batch (line 698) | fn create_record_batch(num_rows: usize) -> RecordBatch { function test_update_batch_with_filter (line 724) | fn test_update_batch_with_filter() { function test_update_batch_filter_null_treated_as_exclude (line 751) | fn test_update_batch_filter_null_treated_as_exclude() { FILE: native/spark-expr/src/agg_funcs/sum_int.rs type SumInteger (line 35) | pub struct SumInteger { method try_new (line 41) | pub fn try_new(data_type: DataType, eval_mode: EvalMode) -> DFResult &dyn Any { method name (line 59) | fn name(&self) -> &str { method signature (line 63) | fn signature(&self) -> &Signature { method return_type (line 67) | fn return_type(&self, _arg_types: &[DataType]) -> DFResult { method accumulator (line 71) | fn accumulator(&self, _acc_args: AccumulatorArgs) -> DFResult DFResult> { method groups_accumulator_supported (line 90) | fn groups_accumulator_supported(&self, _args: AccumulatorArgs) -> bool { method create_groups_accumulator (line 94) | fn create_groups_accumulator( method reverse_expr (line 105) | fn reverse_expr(&self) -> ReversedUDAF { type SumIntegerAccumulatorLegacy (line 111) | struct SumIntegerAccumulatorLegacy { method new (line 116) | fn new() -> Self { method update_batch (line 122) | fn update_batch(&mut self, values: &[ArrayRef]) -> DFResult<()> { method evaluate (line 163) | fn evaluate(&mut self) -> DFResult { method size (line 167) | fn size(&self) -> usize { method state (line 171) | fn state(&mut self) -> DFResult> { method merge_batch (line 175) | fn merge_batch(&mut self, states: &[ArrayRef]) -> DFResult<()> { type SumIntegerAccumulatorAnsi (line 204) | struct SumIntegerAccumulatorAnsi { method new (line 209) | fn new() -> Self { method update_batch (line 215) | fn update_batch(&mut self, values: &[ArrayRef]) -> DFResult<()> { method evaluate (line 258) | fn evaluate(&mut self) -> DFResult { method size (line 262) | fn size(&self) -> usize { method state (line 266) | fn state(&mut self) -> DFResult> { method merge_batch (line 270) | fn merge_batch(&mut self, states: &[ArrayRef]) -> DFResult<()> { type SumIntegerAccumulatorTry (line 304) | struct SumIntegerAccumulatorTry { method new (line 310) | fn new() -> Self { method overflowed (line 318) | fn overflowed(&self) -> bool { method update_batch (line 324) | fn update_batch(&mut self, values: &[ArrayRef]) -> DFResult<()> { method evaluate (line 375) | fn evaluate(&mut self) -> DFResult { method size (line 383) | fn size(&self) -> usize { method state (line 387) | fn state(&mut self) -> DFResult> { method merge_batch (line 394) | fn merge_batch(&mut self, states: &[ArrayRef]) -> DFResult<()> { type SumIntGroupsAccumulatorLegacy (line 438) | struct SumIntGroupsAccumulatorLegacy { method new (line 443) | fn new() -> Self { method update_batch (line 449) | fn update_batch( method evaluate (line 520) | fn evaluate(&mut self, emit_to: EmitTo) -> DFResult { method state (line 534) | fn state(&mut self, emit_to: EmitTo) -> DFResult> { method merge_batch (line 539) | fn merge_batch( method size (line 577) | fn size(&self) -> usize { type SumIntGroupsAccumulatorAnsi (line 582) | struct SumIntGroupsAccumulatorAnsi { method new (line 587) | fn new() -> Self { method update_batch (line 593) | fn update_batch( method evaluate (line 667) | fn evaluate(&mut self, emit_to: EmitTo) -> DFResult { method state (line 681) | fn state(&mut self, emit_to: EmitTo) -> DFResult> { method merge_batch (line 686) | fn merge_batch( method size (line 728) | fn size(&self) -> usize { type SumIntGroupsAccumulatorTry (line 733) | struct SumIntGroupsAccumulatorTry { method new (line 739) | fn new() -> Self { method group_overflowed (line 746) | fn group_overflowed(&self, group_index: usize) -> bool { method update_batch (line 752) | fn update_batch( method evaluate (line 836) | fn evaluate(&mut self, emit_to: EmitTo) -> DFResult { method state (line 861) | fn state(&mut self, emit_to: EmitTo) -> DFResult> { method merge_batch (line 870) | fn merge_batch( method size (line 934) | fn size(&self) -> usize { function run_update_batch_with_filter (line 945) | fn run_update_batch_with_filter( function test_legacy_update_batch_with_filter (line 964) | fn test_legacy_update_batch_with_filter() { function test_legacy_update_batch_filter_null_treated_as_exclude (line 978) | fn test_legacy_update_batch_filter_null_treated_as_exclude() { function test_ansi_update_batch_with_filter (line 995) | fn test_ansi_update_batch_with_filter() { function test_try_update_batch_with_filter (line 1009) | fn test_try_update_batch_with_filter() { function test_no_filter_still_works (line 1022) | fn test_no_filter_still_works() { FILE: native/spark-expr/src/agg_funcs/variance.rs type Variance (line 38) | pub struct Variance { method hash (line 46) | fn hash(&self, state: &mut H) { method new (line 56) | pub fn new( method as_any (line 75) | fn as_any(&self) -> &dyn Any { method name (line 79) | fn name(&self) -> &str { method signature (line 83) | fn signature(&self) -> &Signature { method return_type (line 87) | fn return_type(&self, _arg_types: &[DataType]) -> Result { method accumulator (line 91) | fn accumulator(&self, _acc_args: AccumulatorArgs) -> Result Result Result> { method default_value (line 125) | fn default_value(&self, _data_type: &DataType) -> Result { type VarianceAccumulator (line 132) | pub struct VarianceAccumulator { method try_new (line 142) | pub fn try_new(s_type: StatsType, null_on_divide_by_zero: bool) -> Res... method get_count (line 152) | pub fn get_count(&self) -> f64 { method get_mean (line 156) | pub fn get_mean(&self) -> f64 { method get_m2 (line 160) | pub fn get_m2(&self) -> f64 { method state (line 166) | fn state(&mut self) -> Result> { method update_batch (line 174) | fn update_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method retract_batch (line 192) | fn retract_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method merge_batch (line 210) | fn merge_batch(&mut self, states: &[ArrayRef]) -> Result<()> { method evaluate (line 232) | fn evaluate(&mut self) -> Result { method size (line 257) | fn size(&self) -> usize { FILE: native/spark-expr/src/array_funcs/array_compact.rs type SparkArrayCompact (line 40) | pub struct SparkArrayCompact { method new (line 51) | pub fn new() -> Self { method default (line 45) | fn default() -> Self { method as_any (line 59) | fn as_any(&self) -> &dyn Any { method name (line 63) | fn name(&self) -> &str { method signature (line 67) | fn signature(&self) -> &Signature { method return_type (line 71) | fn return_type(&self, _arg_types: &[DataType]) -> Result { method return_field_from_args (line 75) | fn return_field_from_args( method invoke_with_args (line 82) | fn invoke_with_args(&self, args: ScalarFunctionArgs) -> Result( FILE: native/spark-expr/src/array_funcs/array_insert.rs constant MAX_ROUNDED_ARRAY_LENGTH (line 41) | const MAX_ROUNDED_ARRAY_LENGTH: usize = 2147483632; type ArrayInsert (line 44) | pub struct ArrayInsert { method new (line 69) | pub fn new( method array_type (line 83) | pub fn array_type(&self, data_type: &DataType) -> DataFusionResult(&self, state: &mut H) { method eq (line 60) | fn eq(&self, other: &Self) -> bool { method as_any (line 95) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 99) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 103) | fn data_type(&self, input_schema: &Schema) -> DataFusionResult { method nullable (line 107) | fn nullable(&self, input_schema: &Schema) -> DataFusionResult { method evaluate (line 111) | fn evaluate(&self, batch: &RecordBatch) -> DataFusionResult Vec<&Arc> { method with_new_children (line 178) | fn with_new_children( function array_insert (line 194) | fn array_insert( method fmt (line 328) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function test_array_insert (line 347) | fn test_array_insert() -> Result<()> { function test_array_insert_negative_index (line 394) | fn test_array_insert_negative_index() -> Result<()> { function test_array_insert_legacy_mode (line 430) | fn test_array_insert_legacy_mode() -> Result<()> { function test_array_insert_bug_repro_null_item_pos1_fixed (line 463) | fn test_array_insert_bug_repro_null_item_pos1_fixed() -> Result<()> { FILE: native/spark-expr/src/array_funcs/get_array_struct_fields.rs type GetArrayStructFields (line 36) | pub struct GetArrayStructFields { method new (line 54) | pub fn new(child: Arc, ordinal: usize) -> Self { method list_field (line 58) | fn list_field(&self, input_schema: &Schema) -> DataFusionResult DataFusionResult(&self, state: &mut H) { method eq (line 48) | fn eq(&self, other: &Self) -> bool { method as_any (line 78) | fn as_any(&self) -> &dyn Any { method data_type (line 82) | fn data_type(&self, input_schema: &Schema) -> DataFusionResult { method nullable (line 93) | fn nullable(&self, input_schema: &Schema) -> DataFusionResult { method evaluate (line 98) | fn evaluate(&self, batch: &RecordBatch) -> DataFusionResult Vec<&Arc> { method with_new_children (line 122) | fn with_new_children( method fmt_sql (line 135) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function get_array_struct_fields (line 140) | fn get_array_struct_fields( method fmt (line 183) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { FILE: native/spark-expr/src/array_funcs/list_extract.rs type ListExtract (line 37) | pub struct ListExtract { method new (line 74) | pub fn new( method child_field (line 94) | fn child_field(&self, input_schema: &Schema) -> DataFusionResult DataFusionError { method hash (line 48) | fn hash(&self, state: &mut H) { method eq (line 60) | fn eq(&self, other: &Self) -> bool { method as_any (line 116) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 120) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 124) | fn data_type(&self, input_schema: &Schema) -> DataFusionResult { method nullable (line 128) | fn nullable(&self, input_schema: &Schema) -> DataFusionResult { method evaluate (line 133) | fn evaluate(&self, batch: &RecordBatch) -> DataFusionResult Vec<&Arc> { method with_new_children (line 205) | fn with_new_children( function one_based_index (line 224) | fn one_based_index( function zero_based_index (line 245) | fn zero_based_index( function list_extract (line 262) | fn list_extract( method fmt (line 320) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function test_list_extract_default_value (line 338) | fn test_list_extract_default_value() -> Result<()> { function test_list_extract_null_index (line 392) | fn test_list_extract_null_index() -> Result<()> { FILE: native/spark-expr/src/array_funcs/size.rs function spark_size (line 29) | pub fn spark_size(args: &[ColumnarValue]) -> Result Self { method default (line 52) | fn default() -> Self { method as_any (line 76) | fn as_any(&self) -> &dyn Any { method name (line 80) | fn name(&self) -> &str { method signature (line 84) | fn signature(&self) -> &Signature { method return_type (line 88) | fn return_type(&self, _arg_types: &[DataType]) -> DataFusionResult DataFusionResult... function spark_size_array (line 97) | fn spark_size_array(array: &ArrayRef) -> Result Result) -> i32 { method as_any (line 78) | fn as_any(&self) -> &dyn Any { method name (line 82) | fn name(&self) -> &str { method signature (line 86) | fn signature(&self) -> &Signature { method return_type (line 90) | fn return_type(&self, _arg_types: &[DataType]) -> Result { method accumulator (line 94) | fn accumulator(&self, _acc_args: AccumulatorArgs) -> Result Result> { method groups_accumulator_supported (line 105) | fn groups_accumulator_supported(&self, _args: AccumulatorArgs) -> bool { method update_batch (line 111) | fn update_batch(&mut self, values: &[ArrayRef]) -> Result<()> { method evaluate (line 143) | fn evaluate(&mut self) -> Result { method size (line 147) | fn size(&self) -> usize { method state (line 151) | fn state(&mut self) -> Result> { method merge_batch (line 158) | fn merge_batch(&mut self, states: &[ArrayRef]) -> Result<()> { FILE: native/spark-expr/src/bloom_filter/bloom_filter_might_contain.rs type BloomFilterMightContain (line 31) | pub struct BloomFilterMightContain { method try_new (line 37) | pub fn try_new(bloom_filter_expr: Arc) -> Result &dyn Any { method name (line 70) | fn name(&self) -> &str { method signature (line 74) | fn signature(&self) -> &Signature { method return_type (line 78) | fn return_type(&self, _arg_types: &[DataType]) -> Result { method invoke_with_args (line 82) | fn invoke_with_args(&self, args: ScalarFunctionArgs) -> Result>>(result: ColumnarValue, expect... function assert_all_null (line 134) | fn assert_all_null(result: ColumnarValue) { function test_execute_scalar_contained (line 141) | fn test_execute_scalar_contained() { function test_execute_scalar_not_contained (line 154) | fn test_execute_scalar_not_contained() { function test_execute_scalar_null (line 167) | fn test_execute_scalar_null() { function test_execute_array (line 180) | fn test_execute_array() { function test_execute_array_partially_null (line 195) | fn test_execute_array_partially_null() { function test_execute_scalar_missing_filter (line 210) | fn test_execute_scalar_missing_filter() { function test_execute_array_missing_filter (line 218) | fn test_execute_array_missing_filter() { FILE: native/spark-expr/src/bloom_filter/spark_bit_array.rs type SparkBitArray (line 25) | pub struct SparkBitArray { method new (line 31) | pub fn new(buf: Vec) -> Self { method set (line 39) | pub fn set(&mut self, index: usize) -> bool { method get (line 50) | pub fn get(&self, index: usize) -> bool { method bit_size (line 60) | pub fn bit_size(&self) -> u64 { method byte_size (line 64) | pub fn byte_size(&self) -> usize { method word_size (line 68) | pub fn word_size(&self) -> usize { method cardinality (line 73) | pub fn cardinality(&self) -> usize { method to_bytes (line 77) | pub fn to_bytes(&self) -> Vec { method data (line 81) | pub fn data(&self) -> Vec { method merge_bits (line 87) | pub fn merge_bits(&mut self, other: &[u8]) { function num_words (line 104) | pub fn num_words(num_bits: usize) -> usize { function test_spark_bit_array (line 113) | fn test_spark_bit_array() { function test_spark_bit_with_non_empty_buffer (line 150) | fn test_spark_bit_with_non_empty_buffer() { function test_spark_bit_with_empty_buffer (line 175) | fn test_spark_bit_with_empty_buffer() { function test_spark_bit_with_full_buffer (line 188) | fn test_spark_bit_with_full_buffer() { function test_spark_bit_merge (line 201) | fn test_spark_bit_merge() { function test_num_words (line 238) | fn test_num_words() { FILE: native/spark-expr/src/bloom_filter/spark_bloom_filter.rs constant SPARK_BLOOM_FILTER_VERSION_1 (line 26) | const SPARK_BLOOM_FILTER_VERSION_1: i32 = 1; type SparkBloomFilter (line 32) | pub struct SparkBloomFilter { method from (line 46) | fn from((num_hash_functions, num_bits): (i32, i32)) -> Self { method from (line 59) | fn from(buf: &[u8]) -> Self { method spark_serialization (line 86) | pub fn spark_serialization(&self) -> Vec { method put_long (line 101) | pub fn put_long(&mut self, item: i64) -> bool { method put_binary (line 118) | pub fn put_binary(&mut self, item: &[u8]) -> bool { method might_contain_long (line 135) | pub fn might_contain_long(&self, item: i64) -> bool { method might_contain_longs (line 151) | pub fn might_contain_longs(&self, items: &Int64Array) -> BooleanArray { method state_as_bytes (line 158) | pub fn state_as_bytes(&self) -> Vec { method merge_filter (line 162) | pub fn merge_filter(&mut self, other: &[u8]) { function optimal_num_hash_functions (line 37) | pub fn optimal_num_hash_functions(expected_items: i32, num_bits: i32) ->... FILE: native/spark-expr/src/comet_scalar_funcs.rs function create_comet_physical_fun (line 81) | pub fn create_comet_physical_fun( function create_comet_physical_fun_with_eval_mode (line 97) | pub fn create_comet_physical_fun_with_eval_mode( function all_scalar_functions (line 201) | fn all_scalar_functions() -> Vec> { function register_all_comet_functions (line 214) | pub fn register_all_comet_functions(registry: &mut dyn FunctionRegistry)... type CometScalarFunction (line 223) | struct CometScalarFunction { method hash (line 242) | fn hash(&self, state: &mut H) { method new (line 261) | fn new( method eq (line 231) | fn eq(&self, other: &Self) -> bool { method fmt (line 251) | fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { method as_any (line 277) | fn as_any(&self) -> &dyn Any { method name (line 281) | fn name(&self) -> &str { method signature (line 285) | fn signature(&self) -> &Signature { method return_type (line 289) | fn return_type(&self, _: &[DataType]) -> DataFusionResult { method invoke_with_args (line 293) | fn invoke_with_args(&self, args: ScalarFunctionArgs) -> DataFusionResult... FILE: native/spark-expr/src/conditional_funcs/if_expr.rs type IfExpr (line 32) | pub struct IfExpr { method fmt (line 58) | fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { method new (line 69) | pub fn new( method hash (line 41) | fn hash(&self, state: &mut H) { method eq (line 49) | fn eq(&self, other: &Self) -> bool { method as_any (line 87) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 91) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 95) | fn data_type(&self, input_schema: &Schema) -> Result { method nullable (line 100) | fn nullable(&self, _input_schema: &Schema) -> Result { method evaluate (line 108) | fn evaluate(&self, batch: &RecordBatch) -> Result { method children (line 112) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 116) | fn with_new_children( function if_fn (line 139) | fn if_fn( function test_if_1 (line 148) | fn test_if_1() -> Result<()> { function test_if_2 (line 176) | fn test_if_2() -> Result<()> { function test_if_children (line 198) | fn test_if_children() { FILE: native/spark-expr/src/conversion_funcs/boolean.rs function is_df_cast_from_bool_spark_compatible (line 23) | pub fn is_df_cast_from_bool_spark_compatible(to_type: &DataType) -> bool { function cast_boolean_to_decimal (line 31) | pub fn cast_boolean_to_decimal( function cast_boolean_to_timestamp (line 60) | pub(crate) fn cast_boolean_to_timestamp( function test_input_bool_array (line 92) | fn test_input_bool_array() -> ArrayRef { function test_input_spark_opts (line 96) | fn test_input_spark_opts() -> SparkCastOptions { function test_is_df_cast_from_bool_spark_compatible (line 101) | fn test_is_df_cast_from_bool_spark_compatible() { function test_bool_to_int8_cast (line 117) | fn test_bool_to_int8_cast() { function test_bool_to_int16_cast (line 131) | fn test_bool_to_int16_cast() { function test_bool_to_int32_cast (line 145) | fn test_bool_to_int32_cast() { function test_bool_to_int64_cast (line 159) | fn test_bool_to_int64_cast() { function test_bool_to_float32_cast (line 173) | fn test_bool_to_float32_cast() { function test_bool_to_float64_cast (line 187) | fn test_bool_to_float64_cast() { function test_bool_to_string_cast (line 201) | fn test_bool_to_string_cast() { function test_bool_to_decimal_cast (line 215) | fn test_bool_to_decimal_cast() { function test_cast_boolean_to_timestamp (line 232) | fn test_cast_boolean_to_timestamp() { FILE: native/spark-expr/src/conversion_funcs/cast.rs type Cast (line 81) | pub struct Cast { method new (line 106) | pub fn new( method eq (line 90) | fn eq(&self, other: &Self) -> bool { method hash (line 98) | fn hash(&self, state: &mut H) { type SparkCastOptions (line 125) | pub struct SparkCastOptions { method new (line 149) | pub fn new(eval_mode: EvalMode, timezone: &str, allow_incompat: bool) ... method new_without_timezone (line 162) | pub fn new_without_timezone(eval_mode: EvalMode, allow_incompat: bool)... method new_with_version (line 175) | pub fn new_with_version( function spark_cast (line 191) | pub fn spark_cast( function dict_from_values (line 216) | fn dict_from_values( function cast_array (line 249) | pub(crate) fn cast_array( function is_datafusion_spark_compatible (line 465) | fn is_datafusion_spark_compatible(from_type: &DataType, to_type: &DataTy... function cast_struct_to_struct (line 495) | fn cast_struct_to_struct( function cast_map_to_map (line 532) | fn cast_map_to_map( function cast_array_to_string (line 620) | fn cast_array_to_string( function casts_struct_to_string (line 666) | fn casts_struct_to_string( method fmt (line 714) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 724) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 728) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 732) | fn data_type(&self, _: &Schema) -> DataFusionResult { method nullable (line 736) | fn nullable(&self, _: &Schema) -> DataFusionResult { method evaluate (line 740) | fn evaluate(&self, batch: &RecordBatch) -> DataFusionResult Vec<&Arc> { method with_new_children (line 765) | fn with_new_children( function cast_binary_to_string (line 782) | fn cast_binary_to_string( function spark_binary_formatter (line 812) | fn spark_binary_formatter(value: &[u8], binary_output_style: BinaryOutpu... function cast_binary_formatter (line 843) | fn cast_binary_formatter(value: &[u8]) -> String { function test_cast_unsupported_timestamp_to_date (line 858) | fn test_cast_unsupported_timestamp_to_date() { function test_cast_invalid_timezone (line 871) | fn test_cast_invalid_timezone() { function test_cast_struct_to_utf8 (line 883) | fn test_cast_struct_to_utf8() { function test_cast_struct_to_struct (line 912) | fn test_cast_struct_to_struct() { function test_cast_struct_to_struct_drop_column (line 946) | fn test_cast_struct_to_struct_drop_column() { function test_cast_string_array_to_string (line 979) | fn test_cast_string_array_to_string() { function test_cast_i32_array_to_string (line 1003) | fn test_cast_i32_array_to_string() { function test_cast_array_of_nulls_to_array (line 1026) | fn test_cast_array_of_nulls_to_array() { FILE: native/spark-expr/src/conversion_funcs/numeric.rs function is_df_cast_from_int_spark_compatible (line 34) | pub(crate) fn is_df_cast_from_int_spark_compatible(to_type: &DataType) -... function is_df_cast_from_float_spark_compatible (line 49) | pub(crate) fn is_df_cast_from_float_spark_compatible(to_type: &DataType)... function is_df_cast_from_decimal_spark_compatible (line 63) | pub(crate) fn is_df_cast_from_decimal_spark_compatible(to_type: &DataTyp... function format_decimal_str (line 557) | pub(crate) fn format_decimal_str(value_str: &str, precision: usize, scal... function cast_decimal128_to_utf8 (line 586) | pub(crate) fn cast_decimal128_to_utf8(array: &ArrayRef, scale: i8) -> Sp... function decimal128_to_java_string (line 612) | fn decimal128_to_java_string(unscaled: i128, scale: i8, out: &mut String) { function spark_cast_float64_to_utf8 (line 646) | pub(crate) fn spark_cast_float64_to_utf8( function spark_cast_float32_to_utf8 (line 656) | pub(crate) fn spark_cast_float32_to_utf8( function cast_int_to_decimal128_internal (line 666) | fn cast_int_to_decimal128_internal( function cast_int_to_decimal128 (line 720) | pub(crate) fn cast_int_to_decimal128( function spark_cast_int_to_int (line 768) | pub(crate) fn spark_cast_int_to_int( function spark_cast_decimal_to_boolean (line 800) | pub(crate) fn spark_cast_decimal_to_boolean(array: &dyn Array) -> SparkR... function cast_float64_to_decimal128 (line 813) | pub(crate) fn cast_float64_to_decimal128( function cast_float32_to_decimal128 (line 822) | pub(crate) fn cast_float32_to_decimal128( function cast_floating_point_to_decimal128 (line 831) | fn cast_floating_point_to_decimal128( function spark_cast_nonintegral_numeric_to_integral (line 877) | pub(crate) fn spark_cast_nonintegral_numeric_to_integral( function cast_int_to_timestamp (line 1017) | pub(crate) fn cast_int_to_timestamp( function cast_decimal_to_timestamp (line 1040) | pub(crate) fn cast_decimal_to_timestamp( function cast_float_to_timestamp (line 1066) | pub(crate) fn cast_float_to_timestamp( function test_spark_cast_int_to_int_overflow (line 1099) | fn test_spark_cast_int_to_int_overflow() { function test_spark_cast_decimal_to_boolean (line 1121) | fn test_spark_cast_decimal_to_boolean() { function test_cast_int_to_decimal128 (line 1136) | fn test_cast_int_to_decimal128() { function test_cast_int_to_timestamp (line 1153) | fn test_cast_int_to_timestamp() { function test_cast_float_to_decimal (line 1249) | fn test_cast_float_to_decimal() { function test_cast_decimal_to_timestamp (line 1280) | fn test_cast_decimal_to_timestamp() { function test_cast_float_to_timestamp (line 1330) | fn test_cast_float_to_timestamp() { function test_decimal128_to_java_string (line 1389) | fn test_decimal128_to_java_string() { FILE: native/spark-expr/src/conversion_funcs/string.rs type TimeStampInfo (line 115) | struct TimeStampInfo { method with_year (line 140) | fn with_year(&mut self, year: i32) -> &mut Self { method with_month (line 145) | fn with_month(&mut self, month: u32) -> &mut Self { method with_day (line 150) | fn with_day(&mut self, day: u32) -> &mut Self { method with_hour (line 155) | fn with_hour(&mut self, hour: u32) -> &mut Self { method with_minute (line 160) | fn with_minute(&mut self, minute: u32) -> &mut Self { method with_second (line 165) | fn with_second(&mut self, second: u32) -> &mut Self { method with_microsecond (line 170) | fn with_microsecond(&mut self, microsecond: u32) -> &mut Self { method default (line 126) | fn default() -> Self { function is_df_cast_from_string_spark_compatible (line 176) | pub(crate) fn is_df_cast_from_string_spark_compatible(to_type: &DataType... function cast_string_to_float (line 180) | pub(crate) fn cast_string_to_float( function cast_string_to_float_impl (line 195) | fn cast_string_to_float_impl( function parse_string_to_float (line 231) | fn parse_string_to_float(s: &str) -> Option function spark_cast_utf8_to_boolean (line 260) | pub(crate) fn spark_cast_utf8_to_boolean( function cast_string_to_decimal (line 292) | pub(crate) fn cast_string_to_decimal( function cast_string_to_decimal128_impl (line 313) | fn cast_string_to_decimal128_impl( function cast_string_to_decimal256_impl (line 376) | fn cast_string_to_decimal256_impl( function normalize_fullwidth_digits (line 450) | fn normalize_fullwidth_digits(s: &str) -> String { function parse_string_to_decimal (line 478) | fn parse_string_to_decimal(input_str: &str, precision: u8, scale: i8) ->... function invalid_decimal_cast (line 599) | fn invalid_decimal_cast(value: &str, precision: u8, scale: i8) -> SparkE... function parse_decimal_str (line 609) | fn parse_decimal_str( function cast_string_to_date (line 698) | pub(crate) fn cast_string_to_date( function cast_string_to_timestamp (line 735) | pub(crate) fn cast_string_to_timestamp( function cast_string_to_int (line 768) | pub(crate) fn cast_string_to_int( function finalize_int_result (line 846) | fn finalize_int_result(result: T, negati... function do_parse_string_to_int_legacy (line 857) | fn do_parse_string_to_int_legacy SparkResult> { function parse_string_to_i8_ansi (line 985) | fn parse_string_to_i8_ansi(str: &str) -> SparkResult> { function parse_string_to_i8_try (line 992) | fn parse_string_to_i8_try(str: &str) -> SparkResult> { function parse_string_to_i16_legacy (line 999) | fn parse_string_to_i16_legacy(str: &str) -> SparkResult> { function parse_string_to_i16_ansi (line 1006) | fn parse_string_to_i16_ansi(str: &str) -> SparkResult> { function parse_string_to_i16_try (line 1013) | fn parse_string_to_i16_try(str: &str) -> SparkResult> { function parse_sign (line 1022) | fn parse_sign(bytes: &[u8]) -> Option<(bool, &[u8])> { function invalid_value (line 1032) | pub fn invalid_value(value: &str, from_type: &str, to_type: &str) -> Spa... function get_timestamp_values (line 1040) | fn get_timestamp_values( function days_from_civil (line 1129) | fn days_from_civil(y: i64, m: i64, d: i64) -> i64 { function parse_timestamp_to_micros (line 1138) | fn parse_timestamp_to_micros( function parse_str_to_year_timestamp (line 1256) | fn parse_str_to_year_timestamp(value: &str, tz: &T) -> Spar... function parse_str_to_month_timestamp (line 1260) | fn parse_str_to_month_timestamp(value: &str, tz: &T) -> Spa... function parse_str_to_day_timestamp (line 1264) | fn parse_str_to_day_timestamp(value: &str, tz: &T) -> Spark... function parse_str_to_hour_timestamp (line 1268) | fn parse_str_to_hour_timestamp(value: &str, tz: &T) -> Spar... function parse_str_to_minute_timestamp (line 1272) | fn parse_str_to_minute_timestamp(value: &str, tz: &T) -> Sp... function parse_str_to_second_timestamp (line 1276) | fn parse_str_to_second_timestamp(value: &str, tz: &T) -> Sp... function parse_str_to_microsecond_timestamp (line 1280) | fn parse_str_to_microsecond_timestamp( function timestamp_parser (line 1287) | fn timestamp_parser( function parse_sign_offset (line 1374) | fn parse_sign_offset(s: &str) -> Option { function tz_from_offset_secs (line 1414) | fn tz_from_offset_secs(secs: i32) -> Option { function rfind_str (line 1423) | fn rfind_str(haystack: &str, needle: &str) -> Option { function extract_offset_suffix (line 1447) | fn extract_offset_suffix(value: &str) -> Option<(&str, timezone::Tz)> { type TimestampParsePattern (line 1517) | type TimestampParsePattern = (&'static Regex, fn(&str, &T) -> SparkRe... function timestamp_parser_with_tz (line 1543) | fn timestamp_parser_with_tz( function parse_str_to_time_only_timestamp (line 1609) | fn parse_str_to_time_only_timestamp(value: &str, tz: &T) ->... function date_parser (line 1650) | fn date_parser(date_str: &str, eval_mode: EvalMode) -> SparkResult SparkResult DataFusionResult<()> { function extreme_year_boundary_test (line 1923) | fn extreme_year_boundary_test() { function test_leading_whitespace_t_hm (line 1948) | fn test_leading_whitespace_t_hm() { function plus_sign_year_test (line 1983) | fn plus_sign_year_test() { function timestamp_parser_test (line 2002) | fn timestamp_parser_test() { function timestamp_parser_fraction_scaling_test (line 2148) | fn timestamp_parser_fraction_scaling_test() { function timestamp_parser_tz_offset_formats_test (line 2192) | fn timestamp_parser_tz_offset_formats_test() { function timestamp_parser_dst_test (line 2328) | fn timestamp_parser_dst_test() { function date_parser_test (line 2359) | fn date_parser_test() { function test_cast_string_to_date (line 2423) | fn test_cast_string_to_date() { function test_cast_string_array_with_valid_dates (line 2444) | fn test_cast_string_array_with_valid_dates() { function test_cast_string_array_with_invalid_dates (line 2472) | fn test_cast_string_array_with_invalid_dates() { function test_cast_string_as_i8 (line 2524) | fn test_cast_string_as_i8() { FILE: native/spark-expr/src/conversion_funcs/temporal.rs function is_df_cast_from_date_spark_compatible (line 25) | pub(crate) fn is_df_cast_from_date_spark_compatible(to_type: &DataType) ... function is_df_cast_from_timestamp_spark_compatible (line 29) | pub(crate) fn is_df_cast_from_timestamp_spark_compatible(to_type: &DataT... function cast_date_to_timestamp (line 36) | pub(crate) fn cast_date_to_timestamp( function test_cast_date_to_timestamp (line 85) | fn test_cast_date_to_timestamp() { FILE: native/spark-expr/src/conversion_funcs/utils.rs constant MICROS_PER_SECOND (line 29) | pub(crate) const MICROS_PER_SECOND: i64 = 1000000; function unary_dyn (line 32) | pub fn unary_dyn(array: &ArrayRef, op: F) -> Result ArrayRef { function trim_end (line 105) | fn trim_end(s: &str) -> &str { function cast_overflow (line 114) | pub fn cast_overflow(value: &str, from_type: &str, to_type: &str) -> Spa... FILE: native/spark-expr/src/csv_funcs/csv_write_options.rs type CsvWriteOptions (line 21) | pub struct CsvWriteOptions { method new (line 42) | pub fn new( method fmt (line 32) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { FILE: native/spark-expr/src/csv_funcs/to_csv.rs type ToCsv (line 33) | pub struct ToCsv { method new (line 56) | pub fn new( method hash (line 40) | fn hash(&self, state: &mut H) { method eq (line 48) | fn eq(&self, other: &Self) -> bool { method fmt (line 70) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 80) | fn as_any(&self) -> &dyn Any { method data_type (line 84) | fn data_type(&self, _: &Schema) -> Result { method nullable (line 88) | fn nullable(&self, input_schema: &Schema) -> Result { method evaluate (line 92) | fn evaluate(&self, batch: &RecordBatch) -> Result { method children (line 103) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 107) | fn with_new_children( method fmt_sql (line 118) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function to_csv_inner (line 123) | pub fn to_csv_inner( function escape_value (line 216) | fn escape_value(value: &str, quote_char: char, escape_char: char, output... FILE: native/spark-expr/src/datetime_funcs/date_diff.rs type SparkDateDiff (line 31) | pub struct SparkDateDiff { method new (line 37) | pub fn new() -> Self { method default (line 49) | fn default() -> Self { method as_any (line 55) | fn as_any(&self) -> &dyn Any { method name (line 59) | fn name(&self) -> &str { method signature (line 63) | fn signature(&self) -> &Signature { method return_type (line 67) | fn return_type(&self, _: &[DataType]) -> Result { method invoke_with_args (line 71) | fn invoke_with_args(&self, args: ScalarFunctionArgs) -> Result &[String] { FILE: native/spark-expr/src/datetime_funcs/date_from_unix_date.rs type SparkDateFromUnixDate (line 30) | pub struct SparkDateFromUnixDate { method new (line 36) | pub fn new() -> Self { method default (line 45) | fn default() -> Self { method as_any (line 51) | fn as_any(&self) -> &dyn Any { method name (line 55) | fn name(&self) -> &str { method signature (line 59) | fn signature(&self) -> &Signature { method return_type (line 63) | fn return_type(&self, _: &[DataType]) -> Result { method invoke_with_args (line 67) | fn invoke_with_args(&self, args: ScalarFunctionArgs) -> Result &[String] { FILE: native/spark-expr/src/datetime_funcs/date_trunc.rs type SparkDateTrunc (line 30) | pub struct SparkDateTrunc { method new (line 36) | pub fn new() -> Self { method default (line 48) | fn default() -> Self { method as_any (line 54) | fn as_any(&self) -> &dyn Any { method name (line 58) | fn name(&self) -> &str { method signature (line 62) | fn signature(&self) -> &Signature { method return_type (line 66) | fn return_type(&self, _: &[DataType]) -> Result { method invoke_with_args (line 70) | fn invoke_with_args(&self, args: ScalarFunctionArgs) -> Result &[String] { FILE: native/spark-expr/src/datetime_funcs/hours.rs constant MICROS_PER_HOUR (line 37) | const MICROS_PER_HOUR: i64 = 3_600_000_000; type SparkHoursTransform (line 40) | pub struct SparkHoursTransform { method new (line 45) | pub fn new() -> Self { method default (line 53) | fn default() -> Self { method as_any (line 59) | fn as_any(&self) -> &dyn Any { method name (line 63) | fn name(&self) -> &str { method signature (line 67) | fn signature(&self) -> &Signature { method return_type (line 71) | fn return_type(&self, _arg_types: &[DataType]) -> datafusion::common::Re... method invoke_with_args (line 75) | fn invoke_with_args( function test_hours_transform_utc (line 117) | fn test_hours_transform_utc() { function test_hours_transform_ntz (line 142) | fn test_hours_transform_ntz() { function test_hours_transform_negative_epoch (line 165) | fn test_hours_transform_negative_epoch() { function test_hours_transform_null (line 190) | fn test_hours_transform_null() { function test_hours_transform_epoch_zero (line 212) | fn test_hours_transform_epoch_zero() { function test_hours_transform_non_utc_timezone (line 234) | fn test_hours_transform_non_utc_timezone() { function test_hours_transform_ntz_ignores_timezone (line 259) | fn test_hours_transform_ntz_ignores_timezone() { FILE: native/spark-expr/src/datetime_funcs/make_date.rs type SparkMakeDate (line 33) | pub struct SparkMakeDate { method new (line 38) | pub fn new() -> Self { method default (line 47) | fn default() -> Self { function cast_to_int32 (line 53) | fn cast_to_int32(arr: &Arc) -> Result> { function make_date (line 64) | fn make_date(year: i32, month: i32, day: i32) -> Option { method as_any (line 78) | fn as_any(&self) -> &dyn Any { method name (line 82) | fn name(&self) -> &str { method signature (line 86) | fn signature(&self) -> &Signature { method return_type (line 90) | fn return_type(&self, _: &[DataType]) -> Result { method invoke_with_args (line 94) | fn invoke_with_args(&self, args: ScalarFunctionArgs) -> Result(&self, state: &mut H) { method eq (line 56) | fn eq(&self, other: &Self) -> bool { method fmt (line 78) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 88) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 92) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 96) | fn data_type(&self, input_schema: &Schema) -> datafusion::common::Result... method nullable (line 106) | fn nullable(&self, _: &Schema) -> datafusion::common::Result { method evaluate (line 110) | fn evaluate(&self, batch: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method with_new_children (line 156) | fn with_new_children( FILE: native/spark-expr/src/datetime_funcs/unix_timestamp.rs constant MICROS_PER_SECOND (line 29) | const MICROS_PER_SECOND: i64 = 1_000_000; type SparkUnixTimestamp (line 32) | pub struct SparkUnixTimestamp { method new (line 39) | pub fn new(timezone: String) -> Self { method as_any (line 49) | fn as_any(&self) -> &dyn Any { method name (line 53) | fn name(&self) -> &str { method signature (line 57) | fn signature(&self) -> &Signature { method return_type (line 61) | fn return_type(&self, arg_types: &[DataType]) -> datafusion::common::Res... method invoke_with_args (line 70) | fn invoke_with_args( method aliases (line 156) | fn aliases(&self) -> &[String] { function test_unix_timestamp_from_timestamp (line 170) | fn test_unix_timestamp_from_timestamp() { function test_unix_timestamp_from_date (line 195) | fn test_unix_timestamp_from_date() { function test_unix_timestamp_with_nulls (line 220) | fn test_unix_timestamp_with_nulls() { FILE: native/spark-expr/src/hash_funcs/murmur3.rs function spark_murmur3_hash (line 28) | pub fn spark_murmur3_hash(args: &[ColumnarValue]) -> Result>(data: T, seed: u32)... function create_hashes_dictionary (line 139) | fn create_hashes_dictionary( function create_murmur3_hashes (line 178) | pub fn create_murmur3_hashes<'a>( function test_murmur3_hash (line 201) | fn test_murmur3_hash Result>(data: T, seed: u64) -> u64 { function create_xxhash64_hashes_dictionary (line 85) | fn create_xxhash64_hashes_dictionary( function create_xxhash64_hashes (line 124) | fn create_xxhash64_hashes<'a>( function test_xxhash64_hash (line 147) | fn test_xxhash64_hash(&self, state: &mut H) { method new (line 57) | pub fn new(expr: Arc, schema: DataType, timezone: &s... method eq (line 77) | fn eq(&self, other: &dyn Any) -> bool { method eq (line 42) | fn eq(&self, other: &Self) -> bool { method fmt (line 67) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 89) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 93) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 97) | fn data_type(&self, _: &Schema) -> Result { method nullable (line 101) | fn nullable(&self, _input_schema: &Schema) -> Result { method evaluate (line 106) | fn evaluate(&self, batch: &RecordBatch) -> Result { method children (line 114) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 118) | fn with_new_children( function json_string_to_struct (line 132) | fn json_string_to_struct(arr: &Arc, schema: &DataType) -> Res... type FieldBuilder (line 195) | enum FieldBuilder { function create_field_builders (line 209) | fn create_field_builders( function append_null_to_all_builders (line 247) | fn append_null_to_all_builders(builders: &mut [FieldBuilder]) { function append_field_value (line 269) | fn append_field_value( function finish_builder (line 382) | fn finish_builder(builder: FieldBuilder) -> Result { function test_simple_struct (line 412) | fn test_simple_struct() -> Result<()> { function test_all_primitive_types (line 460) | fn test_all_primitive_types() -> Result<()> { function test_empty_and_null_json (line 526) | fn test_empty_and_null_json() -> Result<()> { function test_nested_struct (line 570) | fn test_nested_struct() -> Result<()> { FILE: native/spark-expr/src/json_funcs/to_json.rs type ToJson (line 37) | pub struct ToJson { method new (line 57) | pub fn new(expr: Arc, timezone: &str) -> Self { method eq (line 72) | fn eq(&self, other: &dyn Any) -> bool { method hash (line 45) | fn hash(&self, state: &mut H) { method eq (line 51) | fn eq(&self, other: &Self) -> bool { method fmt (line 66) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 82) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 86) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 90) | fn data_type(&self, _: &Schema) -> Result { method nullable (line 94) | fn nullable(&self, input_schema: &Schema) -> Result { method evaluate (line 98) | fn evaluate(&self, batch: &RecordBatch) -> Result { method children (line 106) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 110) | fn with_new_children( function array_to_json_string (line 123) | fn array_to_json_string(arr: &Arc, timezone: &str) -> Result<... function escape_string (line 136) | fn escape_string(input: &str) -> String { function struct_to_json (line 184) | fn struct_to_json(array: &StructArray, timezone: &str) -> Result Result<()> { function test_nested_struct (line 289) | fn test_nested_struct() -> Result<()> { function create_ints (line 334) | fn create_ints() -> Arc> { function create_bools (line 343) | fn create_bools() -> Arc { function create_strings (line 352) | fn create_strings() -> Arc { FILE: native/spark-expr/src/kernels/strings.rs function substring (line 29) | pub fn substring(array: &dyn Array, start: i64, length: u64) -> Result Option { function trunc_days_to_year (line 61) | fn trunc_days_to_year(days: i32) -> Option { function trunc_days_to_quarter (line 70) | fn trunc_days_to_quarter(days: i32) -> Option { function trunc_days_to_month (line 87) | fn trunc_days_to_month(days: i32) -> Option { function trunc_days_to_week (line 95) | fn trunc_days_to_week(days: i32) -> Option { function as_timestamp_tz_with_op (line 105) | fn as_timestamp_tz_with_op, T: ArrowT... function as_timestamp_tz_with_op_single (line 132) | fn as_timestamp_tz_with_op_single( function as_micros_from_unix_epoch_utc (line 158) | fn as_micros_from_unix_epoch_utc(dt: Option>) -> i64 { function trunc_date_to_year (line 163) | fn trunc_date_to_year(dt: T) -> Option { function quarter_month (line 175) | fn quarter_month(dt: &T) -> u32 { function trunc_date_to_quarter (line 180) | fn trunc_date_to_quarter(dt: T) -> Option { function trunc_date_to_month (line 191) | fn trunc_date_to_month(dt: T) -> Option { function trunc_date_to_week (line 201) | fn trunc_date_to_week(dt: T) -> Option function trunc_date_to_day (line 214) | fn trunc_date_to_day(dt: T) -> Option { function trunc_date_to_hour (line 223) | fn trunc_date_to_hour(dt: T) -> Option { function trunc_date_to_minute (line 231) | fn trunc_date_to_minute(dt: T) -> Option { function trunc_date_to_second (line 238) | fn trunc_date_to_second(dt: T) -> Option { function trunc_date_to_ms (line 243) | fn trunc_date_to_ms(dt: T) -> Option { function trunc_date_to_microsec (line 248) | fn trunc_date_to_microsec(dt: T) -> Option { function date_trunc_dyn (line 259) | pub fn date_trunc_dyn(array: &dyn Array, format: String) -> Result( function date_trunc_date32 (line 310) | fn date_trunc_date32(array: &Date32Array, format: String) -> Result( function date_trunc_array_fmt_dict_plain (line 475) | fn date_trunc_array_fmt_dict_plain( function date_trunc_array_fmt_dict_dict (line 486) | fn date_trunc_array_fmt_dict_dict( function timestamp_trunc_dyn (line 506) | pub(crate) fn timestamp_trunc_dyn( function timestamp_trunc (line 532) | pub(crate) fn timestamp_trunc( function timestamp_trunc_array_fmt_dyn (line 616) | pub(crate) fn timestamp_trunc_array_fmt_dyn( function timestamp_trunc_array_fmt_plain_plain (line 763) | fn timestamp_trunc_array_fmt_plain_plain( function timestamp_trunc_array_fmt_plain_dict (line 774) | fn timestamp_trunc_array_fmt_plain_dict( function timestamp_trunc_array_fmt_dict_plain (line 787) | fn timestamp_trunc_array_fmt_dict_plain( function timestamp_trunc_array_fmt_dict_dict (line 800) | fn timestamp_trunc_array_fmt_dict_dict( function test_date_trunc (line 829) | fn test_date_trunc() { function test_date_trunc_array_fmt_dyn (line 853) | fn test_date_trunc_array_fmt_dyn() { function test_timestamp_trunc (line 968) | fn test_timestamp_trunc() { function test_timestamp_trunc_array_fmt_dyn (line 1008) | fn test_timestamp_trunc_array_fmt_dyn() { FILE: native/spark-expr/src/lib.rs type EvalMode (line 94) | pub enum EvalMode { type BinaryOutputStyle (line 109) | pub enum BinaryOutputStyle { function arithmetic_overflow_error (line 117) | pub(crate) fn arithmetic_overflow_error(from_type: &str) -> SparkError { function decimal_sum_overflow_error (line 123) | pub(crate) fn decimal_sum_overflow_error() -> SparkError { function divide_by_zero_error (line 127) | pub(crate) fn divide_by_zero_error() -> SparkError { FILE: native/spark-expr/src/math_funcs/abs.rs function abs (line 67) | pub fn abs(args: &[ColumnarValue]) -> Result Result<()>>(test_fn: F) { function test_abs_u8_scalar (line 328) | fn test_abs_u8_scalar() { function test_abs_i8_scalar (line 347) | fn test_abs_i8_scalar() { function test_abs_i16_scalar (line 374) | fn test_abs_i16_scalar() { function test_abs_i32_scalar (line 401) | fn test_abs_i32_scalar() { function test_abs_i64_scalar (line 428) | fn test_abs_i64_scalar() { function test_abs_decimal128_scalar (line 455) | fn test_abs_decimal128_scalar() { function test_abs_decimal256_scalar (line 488) | fn test_abs_decimal256_scalar() { function test_abs_i8_array (line 521) | fn test_abs_i8_array() { function test_abs_i16_array (line 552) | fn test_abs_i16_array() { function test_abs_i32_array (line 583) | fn test_abs_i32_array() { function test_abs_i64_array (line 614) | fn test_abs_i64_array() { function test_abs_f32_array (line 645) | fn test_abs_f32_array() { function test_abs_f64_array (line 696) | fn test_abs_f64_array() { function test_abs_decimal128_array (line 728) | fn test_abs_decimal128_array() { function test_abs_decimal256_array (line 761) | fn test_abs_decimal256_array() { function test_abs_u64_array (line 794) | fn test_abs_u64_array() { function test_abs_null_scalars (line 825) | fn test_abs_null_scalars() { FILE: native/spark-expr/src/math_funcs/ceil.rs function spark_ceil (line 29) | pub fn spark_ceil( function decimal_ceil_f (line 79) | fn decimal_ceil_f(scale: &i8) -> impl Fn(i128) -> i128 { function test_ceil_f32_array (line 95) | fn test_ceil_f32_array() -> Result<()> { function test_ceil_f64_array (line 124) | fn test_ceil_f64_array() -> Result<()> { function test_ceil_i64_array (line 153) | fn test_ceil_i64_array() -> Result<()> { function test_ceil_decimal128_array (line 166) | fn test_ceil_decimal128_array() -> Result<()> { function test_ceil_f32_scalar (line 191) | fn test_ceil_f32_scalar() -> Result<()> { function test_ceil_f64_scalar (line 203) | fn test_ceil_f64_scalar() -> Result<()> { function test_ceil_i64_scalar (line 215) | fn test_ceil_i64_scalar() -> Result<()> { function test_ceil_decimal128_scalar (line 227) | fn test_ceil_decimal128_scalar() -> Result<()> { FILE: native/spark-expr/src/math_funcs/checked_arithmetic.rs function try_arithmetic_kernel (line 30) | pub fn try_arithmetic_kernel( function checked_add (line 141) | pub fn checked_add( function checked_sub (line 149) | pub fn checked_sub( function checked_mul (line 157) | pub fn checked_mul( function checked_div (line 165) | pub fn checked_div( function checked_arithmetic_internal (line 173) | fn checked_arithmetic_internal( FILE: native/spark-expr/src/math_funcs/div.rs function spark_decimal_div (line 32) | pub fn spark_decimal_div( function spark_decimal_integral_div (line 40) | pub fn spark_decimal_integral_div( function spark_decimal_div_internal (line 53) | fn spark_decimal_div_internal( FILE: native/spark-expr/src/math_funcs/floor.rs function spark_floor (line 29) | pub fn spark_floor( function decimal_floor_f (line 79) | fn decimal_floor_f(scale: &i8) -> impl Fn(i128) -> i128 { function test_floor_f32_array (line 95) | fn test_floor_f32_array() -> Result<()> { function test_floor_f64_array (line 124) | fn test_floor_f64_array() -> Result<()> { function test_floor_i64_array (line 153) | fn test_floor_i64_array() -> Result<()> { function test_floor_decimal128_array (line 166) | fn test_floor_decimal128_array() -> Result<()> { function test_floor_f32_scalar (line 191) | fn test_floor_f32_scalar() -> Result<()> { function test_floor_f64_scalar (line 203) | fn test_floor_f64_scalar() -> Result<()> { function test_floor_i64_scalar (line 215) | fn test_floor_i64_scalar() -> Result<()> { function test_floor_decimal128_scalar (line 227) | fn test_floor_decimal128_scalar() -> Result<()> { FILE: native/spark-expr/src/math_funcs/internal/checkoverflow.rs type CheckOverflow (line 41) | pub struct CheckOverflow { method new (line 66) | pub fn new( method hash (line 50) | fn hash(&self, state: &mut H) { method eq (line 58) | fn eq(&self, other: &Self) -> bool { method fmt (line 84) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 94) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 98) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 102) | fn data_type(&self, _: &Schema) -> datafusion::common::Result { method nullable (line 106) | fn nullable(&self, _: &Schema) -> datafusion::common::Result { method evaluate (line 110) | fn evaluate(&self, batch: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method with_new_children (line 245) | fn with_new_children( type ScalarChild (line 268) | struct ScalarChild(Option, u8, i8); method fmt (line 271) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 277) | fn as_any(&self) -> &dyn Any { method data_type (line 280) | fn data_type(&self, _: &Schema) -> datafusion::common::Result { method nullable (line 283) | fn nullable(&self, _: &Schema) -> datafusion::common::Result { method evaluate (line 286) | fn evaluate(&self, _: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method with_new_children (line 294) | fn with_new_children( method fmt_sql (line 300) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function empty_batch (line 305) | fn empty_batch() -> RecordBatch { function make_check_overflow (line 310) | fn make_check_overflow( function test_scalar_no_overflow_legacy (line 328) | fn test_scalar_no_overflow_legacy() { function test_scalar_overflow_returns_null_in_legacy_mode (line 339) | fn test_scalar_overflow_returns_null_in_legacy_mode() { function test_scalar_null_passthrough_legacy (line 350) | fn test_scalar_null_passthrough_legacy() { function test_scalar_no_overflow_ansi (line 362) | fn test_scalar_no_overflow_ansi() { function test_scalar_overflow_returns_error_in_ansi_mode (line 373) | fn test_scalar_overflow_returns_error_in_ansi_mode() { function test_scalar_null_passthrough_ansi (line 383) | fn test_scalar_null_passthrough_ansi() { FILE: native/spark-expr/src/math_funcs/internal/decimal_rescale_check.rs type DecimalRescaleCheckOverflow (line 41) | pub struct DecimalRescaleCheckOverflow { method new (line 70) | pub fn new( method hash (line 50) | fn hash(&self, state: &mut H) { method eq (line 60) | fn eq(&self, other: &Self) -> bool { method fmt (line 88) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function precision_bound (line 100) | fn precision_bound(precision: u8) -> i128 { function rescale_and_check (line 112) | fn rescale_and_check( method as_any (line 157) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 161) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 165) | fn data_type(&self, _: &Schema) -> datafusion::common::Result { method nullable (line 172) | fn nullable(&self, _: &Schema) -> datafusion::common::Result { method evaluate (line 176) | fn evaluate(&self, batch: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method with_new_children (line 244) | fn with_new_children( function make_batch (line 272) | fn make_batch(values: Vec>, precision: u8, scale: i8) -> Re... function eval_expr (line 279) | fn eval_expr( function test_scale_up (line 301) | fn test_scale_up() { function test_scale_down_with_half_up_rounding (line 311) | fn test_scale_down_with_half_up_rounding() { function test_same_scale_precision_check_only (line 325) | fn test_same_scale_precision_check_only() { function test_overflow_null_in_legacy_mode (line 335) | fn test_overflow_null_in_legacy_mode() { function test_overflow_error_in_ansi_mode (line 345) | fn test_overflow_error_in_ansi_mode() { function test_null_propagation (line 352) | fn test_null_propagation() { function test_scalar_path (line 362) | fn test_scalar_path() { type ScalarChild (line 386) | struct ScalarChild(Option, u8, i8); method fmt (line 389) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 395) | fn as_any(&self) -> &dyn Any { method data_type (line 398) | fn data_type(&self, _: &Schema) -> datafusion::common::Result { method nullable (line 401) | fn nullable(&self, _: &Schema) -> datafusion::common::Result { method evaluate (line 404) | fn evaluate(&self, _batch: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method with_new_children (line 412) | fn with_new_children( method fmt_sql (line 418) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function test_scalar_null (line 424) | fn test_scalar_null() { function test_scalar_overflow_legacy (line 439) | fn test_scalar_overflow_legacy() { function test_scalar_overflow_ansi_returns_error (line 459) | fn test_scalar_overflow_ansi_returns_error() { function test_large_scale_delta_returns_error (line 475) | fn test_large_scale_delta_returns_error() { FILE: native/spark-expr/src/math_funcs/internal/make_decimal.rs function spark_make_decimal (line 29) | pub fn spark_make_decimal( function long_to_decimal (line 64) | fn long_to_decimal(v: &Option, precision: u8, scale: i8) -> Option<... FILE: native/spark-expr/src/math_funcs/internal/normalize_nan.rs type NormalizeNaNAndZero (line 35) | pub struct NormalizeNaNAndZero { method new (line 54) | pub fn new(data_type: DataType, child: Arc) -> Self { method eq (line 41) | fn eq(&self, other: &Self) -> bool { method hash (line 47) | fn hash(&self, state: &mut H) { method as_any (line 60) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 64) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 68) | fn data_type(&self, input_schema: &Schema) -> datafusion::common::Result... method nullable (line 72) | fn nullable(&self, input_schema: &Schema) -> datafusion::common::Result<... method evaluate (line 76) | fn evaluate(&self, batch: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method with_new_children (line 101) | fn with_new_children( function normalize_float (line 115) | fn normalize_float(v: T) -> T { method fmt (line 126) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { FILE: native/spark-expr/src/math_funcs/internal/unscaled_value.rs function spark_unscaled_value (line 27) | pub fn spark_unscaled_value(args: &[ColumnarValue]) -> DataFusionResult<... FILE: native/spark-expr/src/math_funcs/log.rs function spark_log (line 27) | pub fn spark_log(args: &[ColumnarValue]) -> Result Resu... function create_modulo_expr (line 66) | pub fn create_modulo_expr( function null_if_zero_primitive (line 139) | fn null_if_zero_primitive( function is_primitive_datatype (line 178) | fn is_primitive_datatype(dt: &DataType) -> bool { function create_modulo_scalar_function (line 196) | fn create_modulo_scalar_function( function with_fail_on_error (line 226) | fn with_fail_on_error(test_fn: F) { function verify_result (line 232) | pub fn verify_result( function test_modulo_basic_int (line 291) | fn test_modulo_basic_int() { function test_modulo_basic_decimal (line 324) | fn test_modulo_basic_decimal() { function test_modulo_divide_by_zero_int (line 370) | fn test_modulo_divide_by_zero_int() { function test_division_by_zero_with_complex_int_expr (line 402) | fn test_division_by_zero_with_complex_int_expr() { FILE: native/spark-expr/src/math_funcs/negative.rs function create_negate_expr (line 34) | pub fn create_negate_expr( type NegativeExpr (line 43) | pub struct NegativeExpr { method new (line 82) | pub fn new(arg: Arc, fail_on_error: bool) -> Self { method arg (line 87) | pub fn arg(&self) -> &Arc { method fmt (line 93) | fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { method hash (line 50) | fn hash(&self, state: &mut H) { method eq (line 57) | fn eq(&self, other: &Self) -> bool { method as_any (line 100) | fn as_any(&self) -> &dyn Any { method data_type (line 104) | fn data_type(&self, input_schema: &Schema) -> Result { method nullable (line 108) | fn nullable(&self, input_schema: &Schema) -> Result { method evaluate (line 112) | fn evaluate(&self, batch: &RecordBatch) -> Result { method children (line 194) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 198) | fn with_new_children( method evaluate_bounds (line 211) | fn evaluate_bounds(&self, children: &[&Interval]) -> Result { method propagate_constraints (line 220) | fn propagate_constraints( method get_properties (line 249) | fn get_properties(&self, children: &[ExprProperties]) -> Result) -> std::fmt::Result { FILE: native/spark-expr/src/math_funcs/round.rs function spark_round (line 103) | pub fn spark_round( function decimal_round_f (line 187) | fn decimal_round_f(scale: &i8, point: &i64) -> Box i128> { function test_round_f32_array (line 218) | fn test_round_f32_array() -> Result<()> { function test_round_f64_array (line 236) | fn test_round_f64_array() -> Result<()> { function test_round_f32_scalar (line 254) | fn test_round_f32_scalar() -> Result<()> { function test_round_f64_scalar (line 270) | fn test_round_f64_scalar() -> Result<()> { FILE: native/spark-expr/src/math_funcs/unhex.rs function unhex_digit (line 26) | fn unhex_digit(c: u8) -> Result { function unhex (line 39) | fn unhex(hex_str: &str, result: &mut Vec) -> Result<(), DataFusionEr... function spark_unhex_inner (line 62) | fn spark_unhex_inner( function spark_unhex (line 113) | pub fn spark_unhex(args: &[ColumnarValue]) -> Result Result<(), Box> { function test_partial_error (line 175) | fn test_partial_error() -> Result<(), Box> { function test_unhex_valid (line 201) | fn test_unhex_valid() -> Result<(), Box> { function test_odd_length (line 225) | fn test_odd_length() -> Result<(), Box> { function test_unhex_empty (line 240) | fn test_unhex_empty() { function test_unhex_invalid (line 249) | fn test_unhex_invalid() { FILE: native/spark-expr/src/math_funcs/utils.rs function make_decimal_scalar (line 45) | pub(crate) fn make_decimal_scalar( function make_decimal_array (line 56) | pub(crate) fn make_decimal_array( function get_precision_scale (line 69) | pub(crate) fn get_precision_scale(data_type: &DataType) -> (u8, i8) { FILE: native/spark-expr/src/math_funcs/wide_decimal_binary_expr.rs type WideDecimalOp (line 38) | pub enum WideDecimalOp { method fmt (line 45) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { type WideDecimalBinaryExpr (line 58) | pub struct WideDecimalBinaryExpr { method new (line 90) | pub fn new( method hash (line 68) | fn hash(&self, state: &mut H) { method eq (line 79) | fn eq(&self, other: &Self) -> bool { method fmt (line 110) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function div_round_half_up (line 121) | fn div_round_half_up(value: i256, divisor: i256) -> i256 { constant I256_TEN (line 147) | const I256_TEN: i256 = i256::from_i128(10); function i256_pow10 (line 151) | fn i256_pow10(exp: u32) -> i256 { function max_for_precision (line 163) | fn max_for_precision(precision: u8) -> i256 { method as_any (line 168) | fn as_any(&self) -> &dyn Any { method data_type (line 172) | fn data_type(&self, _input_schema: &Schema) -> Result { method nullable (line 179) | fn nullable(&self, _input_schema: &Schema) -> Result { method evaluate (line 183) | fn evaluate(&self, batch: &RecordBatch) -> Result { method children (line 286) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 290) | fn with_new_children( method fmt_sql (line 310) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function check_overflow_and_convert (line 319) | fn check_overflow_and_convert( function make_batch (line 344) | fn make_batch( function eval_expr (line 367) | fn eval_expr( function test_add_same_scale (line 385) | fn test_add_same_scale() { function test_subtract_same_scale (line 402) | fn test_subtract_same_scale() { function test_add_different_scales (line 418) | fn test_add_different_scales() { function test_multiply_with_scale_reduction (line 434) | fn test_multiply_with_scale_reduction() { function test_multiply_half_up_rounding (line 451) | fn test_multiply_half_up_rounding() { function test_multiply_half_up_rounding_negative (line 469) | fn test_multiply_half_up_rounding_negative() { function test_overflow_legacy_mode_returns_null (line 485) | fn test_overflow_legacy_mode_returns_null() { function test_overflow_ansi_mode_returns_error (line 494) | fn test_overflow_ansi_mode_returns_error() { function test_null_propagation (line 501) | fn test_null_propagation() { function test_zeros (line 510) | fn test_zeros() { function test_max_precision_values (line 518) | fn test_max_precision_values() { function test_add_scale_up_to_output (line 528) | fn test_add_scale_up_to_output() { function test_subtract_scale_up_to_output (line 546) | fn test_subtract_scale_up_to_output() { FILE: native/spark-expr/src/nondetermenistic_funcs/internal/rand_utils.rs function evaluate_batch_for_rand (line 23) | pub fn evaluate_batch_for_rand( type StatefulSeedValueGenerator (line 44) | pub trait StatefulSeedValueGenerator: Sized { method from_init_seed (line 45) | fn from_init_seed(init_seed: i64) -> Self; method from_stored_state (line 47) | fn from_stored_state(stored_state: State) -> Self; method next_value (line 49) | fn next_value(&mut self) -> Value; method get_current_state (line 51) | fn get_current_state(&self) -> State; method from_state_ref (line 53) | fn from_state_ref(state: impl Deref>, init_valu... FILE: native/spark-expr/src/nondetermenistic_funcs/monotonically_increasing_id.rs type MonotonicallyIncreasingId (line 30) | pub struct MonotonicallyIncreasingId { method from_offset (line 36) | pub fn from_offset(offset: i64) -> Self { method from_partition_id (line 43) | pub fn from_partition_id(partition: i32) -> Self { method fmt (line 49) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method eq (line 55) | fn eq(&self, other: &Self) -> bool { method hash (line 63) | fn hash(&self, state: &mut H) { method as_any (line 69) | fn as_any(&self) -> &dyn Any { method evaluate (line 73) | fn evaluate(&self, batch: &RecordBatch) -> Result { method children (line 82) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 86) | fn with_new_children( method fmt_sql (line 93) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 97) | fn data_type(&self, _input_schema: &Schema) -> Result { method nullable (line 101) | fn nullable(&self, _input_schema: &Schema) -> Result { function test_monotonically_increasing_id_single_batch (line 115) | fn test_monotonically_increasing_id_single_batch() -> Result<()> { function test_monotonically_increasing_id_multi_batch (line 128) | fn test_monotonically_increasing_id_multi_batch() -> Result<()> { FILE: native/spark-expr/src/nondetermenistic_funcs/rand.rs constant DOUBLE_UNIT (line 36) | const DOUBLE_UNIT: f64 = 1.1102230246251565e-16; constant SPARK_MURMUR_ARRAY_SEED (line 42) | const SPARK_MURMUR_ARRAY_SEED: u32 = 0x3c074a61; type XorShiftRandom (line 45) | pub(crate) struct XorShiftRandom { method next (line 50) | fn next(&mut self, bits: u8) -> i32 { method next_f64 (line 58) | pub fn next_f64(&mut self) -> f64 { method from_init_seed (line 66) | fn from_init_seed(init_seed: i64) -> Self { method from_stored_state (line 74) | fn from_stored_state(stored_state: i64) -> Self { method next_value (line 78) | fn next_value(&mut self) -> f64 { method get_current_state (line 82) | fn get_current_state(&self) -> i64 { type RandExpr (line 88) | pub struct RandExpr { method new (line 94) | pub fn new(seed: i64) -> Self { method fmt (line 103) | fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { method eq (line 109) | fn eq(&self, other: &Self) -> bool { method hash (line 117) | fn hash(&self, state: &mut H) { method as_any (line 123) | fn as_any(&self) -> &dyn Any { method data_type (line 127) | fn data_type(&self, _input_schema: &Schema) -> Result { method nullable (line 131) | fn nullable(&self, _input_schema: &Schema) -> Result { method evaluate (line 135) | fn evaluate(&self, batch: &RecordBatch) -> Result { method children (line 143) | fn children(&self) -> Vec<&Arc> { method fmt_sql (line 147) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method with_new_children (line 151) | fn with_new_children( function rand (line 159) | pub fn rand(seed: i64) -> Arc { constant SPARK_SEED_42_FIRST_5 (line 170) | const SPARK_SEED_42_FIRST_5: [f64; 5] = [ function test_rand_single_batch (line 179) | fn test_rand_single_batch() -> Result<()> { function test_rand_multi_batch (line 192) | fn test_rand_multi_batch() -> Result<()> { FILE: native/spark-expr/src/nondetermenistic_funcs/randn.rs type XorShiftRandomForGaussian (line 44) | struct XorShiftRandomForGaussian { method next_gaussian (line 50) | pub fn next_gaussian(&mut self) -> f64 { method from_init_seed (line 75) | fn from_init_seed(init_value: i64) -> Self { method from_stored_state (line 82) | fn from_stored_state(stored_state: RandomGaussianState) -> Self { method next_value (line 89) | fn next_value(&mut self) -> f64 { method get_current_state (line 93) | fn get_current_state(&self) -> RandomGaussianState { type RandomGaussianState (line 72) | type RandomGaussianState = (i64, Option); type RandnExpr (line 99) | pub struct RandnExpr { method new (line 105) | pub fn new(seed: i64) -> Self { method fmt (line 114) | fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { method eq (line 120) | fn eq(&self, other: &Self) -> bool { method hash (line 128) | fn hash(&self, state: &mut H) { method as_any (line 134) | fn as_any(&self) -> &dyn Any { method data_type (line 138) | fn data_type(&self, _input_schema: &Schema) -> datafusion::common::Resul... method nullable (line 142) | fn nullable(&self, _input_schema: &Schema) -> datafusion::common::Result... method evaluate (line 146) | fn evaluate(&self, batch: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method fmt_sql (line 158) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method with_new_children (line 162) | fn with_new_children( function randn (line 170) | pub fn randn(seed: i64) -> Arc { constant PRECISION_TOLERANCE (line 181) | const PRECISION_TOLERANCE: f64 = 1e-6; constant SPARK_SEED_42_FIRST_5_GAUSSIAN (line 183) | const SPARK_SEED_42_FIRST_5_GAUSSIAN: [f64; 5] = [ function test_rand_single_batch (line 192) | fn test_rand_single_batch() -> datafusion::common::Result<()> { function test_rand_multi_batch (line 205) | fn test_rand_multi_batch() -> datafusion::common::Result<()> { function assert_eq_with_tolerance (line 236) | fn assert_eq_with_tolerance(left: &Float64Array, right: &Float64Array) { FILE: native/spark-expr/src/predicate_funcs/is_nan.rs function spark_isnan (line 26) | pub fn spark_isnan(args: &[ColumnarValue]) -> Result, pattern: &str) -> Result<... method is_match (line 75) | fn is_match(&self, inputs: &StringArray) -> BooleanArray { method eq (line 50) | fn eq(&self, other: &Self) -> bool { method hash (line 58) | fn hash(&self, state: &mut H) { method fmt (line 95) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 105) | fn as_any(&self) -> &dyn Any { method data_type (line 109) | fn data_type(&self, _input_schema: &Schema) -> Result { method nullable (line 113) | fn nullable(&self, input_schema: &Schema) -> Result { method evaluate (line 117) | fn evaluate(&self, batch: &RecordBatch) -> Result { method children (line 165) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 169) | fn with_new_children( method fmt_sql (line 180) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function test_rlike_scalar_string_variants (line 191) | fn test_rlike_scalar_string_variants() { function test_rlike_scalar_non_string_error (line 223) | fn test_rlike_scalar_non_string_error() { FILE: native/spark-expr/src/static_invoke/char_varchar_utils/read_side_padding.rs constant SPACE (line 28) | const SPACE: &str = " "; function spark_read_side_padding (line 30) | pub fn spark_read_side_padding(args: &[ColumnarValue]) -> Result Result Result( function write_padded_string (line 275) | fn write_padded_string( function write_padding_chars (line 316) | fn write_padding_chars(buffer: &mut String, pad_chars: &[char], count: u... FILE: native/spark-expr/src/string_funcs/contains.rs type SparkContains (line 37) | pub struct SparkContains { method new (line 48) | pub fn new() -> Self { method default (line 42) | fn default() -> Self { method as_any (line 56) | fn as_any(&self) -> &dyn Any { method name (line 60) | fn name(&self) -> &str { method signature (line 64) | fn signature(&self) -> &Signature { method return_type (line 68) | fn return_type(&self, _arg_types: &[DataType]) -> Result { method invoke_with_args (line 72) | fn invoke_with_args(&self, args: ScalarFunctionArgs) -> Result R... function contains_with_arrow_scalar (line 112) | fn contains_with_arrow_scalar( function contains_scalar_scalar (line 143) | fn contains_scalar_scalar( function test_contains_array_scalar (line 187) | fn test_contains_array_scalar() { function test_contains_scalar_scalar (line 206) | fn test_contains_scalar_scalar() { function test_contains_null_needle (line 219) | fn test_contains_null_needle() { function test_contains_empty_needle (line 235) | fn test_contains_empty_needle() { FILE: native/spark-expr/src/string_funcs/get_json_object.rs function scalar_to_str (line 27) | fn scalar_to_str(scalar: &ScalarValue, arg_name: &str) -> DataFusionResu... function spark_get_json_object (line 44) | pub fn spark_get_json_object(args: &[ColumnarValue]) -> DataFusionResult... type PathSegment (line 137) | enum PathSegment { type ParsedPath (line 147) | struct ParsedPath { function parse_json_path (line 154) | fn parse_json_path(path: &str) -> Option { function evaluate_path (line 254) | fn evaluate_path(json_str: &str, path: &ParsedPath) -> Option { function evaluate_no_wildcard (line 288) | fn evaluate_no_wildcard<'a>(value: &'a Value, segments: &[PathSegment]) ... function evaluate_with_wildcard (line 308) | fn evaluate_with_wildcard<'a>(value: &'a Value, segments: &[PathSegment]... function value_to_string (line 344) | fn value_to_string(value: &Value) -> Option { function test_parse_json_path (line 357) | fn test_parse_json_path() { function test_evaluate_simple_field (line 394) | fn test_evaluate_simple_field() { function test_evaluate_nested (line 408) | fn test_evaluate_nested() { function test_evaluate_array_index (line 415) | fn test_evaluate_array_index() { function test_evaluate_root (line 423) | fn test_evaluate_root() { function test_evaluate_null_value (line 432) | fn test_evaluate_null_value() { function test_evaluate_missing_field (line 438) | fn test_evaluate_missing_field() { function test_evaluate_invalid_json (line 444) | fn test_evaluate_invalid_json() { function test_evaluate_wildcard (line 450) | fn test_evaluate_wildcard() { function test_evaluate_string_unquoted (line 457) | fn test_evaluate_string_unquoted() { function test_evaluate_nested_array_field (line 464) | fn test_evaluate_nested_array_field() { function test_evaluate_bracket_notation_with_spaces (line 471) | fn test_evaluate_bracket_notation_with_spaces() { function test_evaluate_boolean_and_nested_object (line 478) | fn test_evaluate_boolean_and_nested_object() { function test_object_key_order_preserved (line 487) | fn test_object_key_order_preserved() { function test_wildcard_single_match (line 498) | fn test_wildcard_single_match() { function test_wildcard_missing_fields (line 510) | fn test_wildcard_missing_fields() { function test_field_with_colon (line 518) | fn test_field_with_colon() { function test_dot_bracket_invalid (line 525) | fn test_dot_bracket_invalid() { function test_object_wildcard (line 531) | fn test_object_wildcard() { function test_unicode_field_names (line 539) | fn test_unicode_field_names() { function test_unicode_values (line 546) | fn test_unicode_values() { function test_unicode_emoji (line 556) | fn test_unicode_emoji() { function test_unicode_bracket_notation (line 565) | fn test_unicode_bracket_notation() { function test_unicode_mixed_scripts (line 572) | fn test_unicode_mixed_scripts() { function test_unicode_wildcard (line 582) | fn test_unicode_wildcard() { FILE: native/spark-expr/src/string_funcs/split.rs function spark_split (line 38) | pub fn spark_split(args: &[ColumnarValue]) -> DataFusionResult DataFusionRe... function split_with_regex (line 211) | fn split_with_regex(string: &str, regex: &Regex, limit: i32) -> Vec ListArray { function new_null_list_array (line 256) | fn new_null_list_array(len: usize) -> ArrayRef { function test_split_basic (line 276) | fn test_split_basic() { function test_split_with_limit (line 287) | fn test_split_with_limit() { function test_split_regex (line 299) | fn test_split_regex() { function test_split_limit_positive (line 305) | fn test_split_limit_positive() { function test_split_limit_zero (line 311) | fn test_split_limit_zero() { function test_split_limit_negative (line 317) | fn test_split_limit_negative() { function test_split_with_nulls (line 323) | fn test_split_with_nulls() { function test_split_empty_string (line 353) | fn test_split_empty_string() { FILE: native/spark-expr/src/string_funcs/substring.rs type SubstringExpr (line 34) | pub struct SubstringExpr { method new (line 55) | pub fn new(child: Arc, start: i64, len: u64) -> Self { method hash (line 41) | fn hash(&self, state: &mut H) { method eq (line 49) | fn eq(&self, other: &Self) -> bool { method fmt (line 61) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method as_any (line 71) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 75) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 79) | fn data_type(&self, input_schema: &Schema) -> datafusion::common::Result... method nullable (line 83) | fn nullable(&self, _: &Schema) -> datafusion::common::Result { method evaluate (line 87) | fn evaluate(&self, batch: &RecordBatch) -> datafusion::common::Result Vec<&Arc> { method with_new_children (line 105) | fn with_new_children( FILE: native/spark-expr/src/struct_funcs/create_named_struct.rs type CreateNamedStruct (line 32) | pub struct CreateNamedStruct { method new (line 38) | pub fn new(values: Vec>, names: Vec) -> ... method fields (line 42) | fn fields(&self, schema: &Schema) -> DataFusionResult> { method as_any (line 56) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 60) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 64) | fn data_type(&self, input_schema: &Schema) -> DataFusionResult { method nullable (line 69) | fn nullable(&self, _input_schema: &Schema) -> DataFusionResult { method evaluate (line 73) | fn evaluate(&self, batch: &RecordBatch) -> DataFusionResult Vec<&Arc> { method with_new_children (line 92) | fn with_new_children( method fmt (line 104) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { function test_create_struct_from_dict_encoded_i32 (line 125) | fn test_create_struct_from_dict_encoded_i32() -> Result<()> { function test_create_struct_from_dict_encoded_string (line 142) | fn test_create_struct_from_dict_encoded_string() -> Result<()> { FILE: native/spark-expr/src/struct_funcs/get_struct_field.rs type GetStructField (line 32) | pub struct GetStructField { method new (line 50) | pub fn new(child: Arc, ordinal: usize) -> Self { method child_field (line 54) | fn child_field(&self, input_schema: &Schema) -> DataFusionResult(&self, state: &mut H) { method eq (line 44) | fn eq(&self, other: &Self) -> bool { method as_any (line 65) | fn as_any(&self) -> &dyn Any { method fmt_sql (line 69) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 73) | fn data_type(&self, input_schema: &Schema) -> DataFusionResult { method nullable (line 77) | fn nullable(&self, input_schema: &Schema) -> DataFusionResult { method evaluate (line 81) | fn evaluate(&self, batch: &RecordBatch) -> DataFusionResult Vec<&Arc> { method with_new_children (line 108) | fn with_new_children( method fmt (line 120) | fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result { FILE: native/spark-expr/src/test_common/file_util.rs function get_temp_file (line 21) | pub fn get_temp_file(file_name: &str, content: &[u8]) -> fs::File { function get_temp_filename (line 44) | pub fn get_temp_filename() -> PathBuf { FILE: native/spark-expr/src/timezone.rs function parse_fixed_offset (line 28) | fn parse_fixed_offset(tz: &str) -> Result { type TzOffset (line 50) | pub struct TzOffset { method fmt (line 56) | fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { method fix (line 62) | fn fix(&self) -> FixedOffset { type Tz (line 69) | pub struct Tz(TzInner); type TzInner (line 72) | enum TzInner { type Err (line 78) | type Err = ArrowError; method from_str (line 80) | fn from_str(tz: &str) -> Result { type Offset (line 101) | type Offset = TzOffset; method from_offset (line 103) | fn from_offset(offset: &Self::Offset) -> Self { method offset_from_local_date (line 107) | fn offset_from_local_date(&self, local: &NaiveDate) -> LocalResult LocalResu... method offset_from_utc_date (line 125) | fn offset_from_utc_date(&self, utc: &NaiveDate) -> Self::Offset { method offset_from_utc_datetime (line 134) | fn offset_from_utc_datetime(&self, utc: &NaiveDateTime) -> Self::Offset { FILE: native/spark-expr/src/unbound.rs type UnboundColumn (line 35) | pub struct UnboundColumn { method new (line 42) | pub fn new(name: &str, datatype: DataType) -> Self { method name (line 50) | pub fn name(&self) -> &str { method fmt (line 56) | fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { method as_any (line 63) | fn as_any(&self) -> &dyn std::any::Any { method fmt_sql (line 67) | fn fmt_sql(&self, f: &mut Formatter<'_>) -> std::fmt::Result { method data_type (line 72) | fn data_type(&self, _input_schema: &Schema) -> Result { method nullable (line 77) | fn nullable(&self, _input_schema: &Schema) -> Result { method evaluate (line 82) | fn evaluate(&self, _batch: &RecordBatch) -> Result { method children (line 86) | fn children(&self) -> Vec<&Arc> { method with_new_children (line 90) | fn with_new_children( FILE: native/spark-expr/src/utils.rs function array_with_timezone (line 69) | pub fn array_with_timezone( function datetime_cast_err (line 171) | fn datetime_cast_err(value: i64) -> ArrowError { function resolve_local_datetime (line 182) | fn resolve_local_datetime(tz: &Tz, local_datetime: NaiveDateTime) -> Dat... function timestamp_ntz_to_timestamp (line 213) | fn timestamp_ntz_to_timestamp( function pre_timestamp_cast (line 263) | fn pre_timestamp_cast(array: ArrayRef, timezone: String) -> Result bool { function build_bool_state (line 303) | pub fn build_bool_state(state: &mut BooleanBufferBuilder, emit_to: &Emit... function cold (line 322) | pub fn cold() {} function likely (line 325) | pub fn likely(b: bool) -> bool { function unlikely (line 332) | pub fn unlikely(b: bool) -> bool { function array_containing (line 343) | fn array_containing(local_datetime: &str) -> ArrayRef { function micros_for (line 349) | fn micros_for(datetime: &str) -> i64 { function test_build_bool_state (line 357) | fn test_build_bool_state() { function test_timestamp_ntz_to_timestamp_handles_non_existent_time (line 376) | fn test_timestamp_ntz_to_timestamp_handles_non_existent_time() { function test_timestamp_ntz_to_timestamp_handles_ambiguous_time (line 391) | fn test_timestamp_ntz_to_timestamp_handles_ambiguous_time() { FILE: native/spark-expr/tests/spark_expr_reg.rs function test_udf_registration (line 29) | async fn test_udf_registration() -> Result<()> { function test_make_date_returns_null_for_invalid_input (line 54) | async fn test_make_date_returns_null_for_invalid_input() -> Result<()> { function test_make_date_valid_input (line 76) | async fn test_make_date_valid_input() -> Result<()> { FILE: spark/src/main/java/org/apache/comet/CometBatchIterator.java class CometBatchIterator (line 37) | public class CometBatchIterator { method CometBatchIterator (line 43) | CometBatchIterator(Iterator input, NativeUtil nativeUti... method hasNext (line 54) | public int hasNext() { method next (line 78) | public int next(long[] arrayAddrs, long[] schemaAddrs) { method hasSelectionVectors (line 100) | public boolean hasSelectionVectors() { method exportSelectionIndices (line 121) | public int exportSelectionIndices(long[] arrayAddrs, long[] schemaAddr... FILE: spark/src/main/java/org/apache/comet/CometShuffleBlockIterator.java class CometShuffleBlockIterator (line 42) | public class CometShuffleBlockIterator implements Closeable { method CometShuffleBlockIterator (line 53) | public CometShuffleBlockIterator(InputStream in) { method hasNext (line 67) | public int hasNext() throws IOException { method getBuffer (line 126) | public ByteBuffer getBuffer() { method getCurrentBlockLength (line 131) | public int getCurrentBlockLength() { method close (line 135) | @Override FILE: spark/src/main/java/org/apache/comet/NativeColumnarToRowInfo.java class NativeColumnarToRowInfo (line 44) | public class NativeColumnarToRowInfo { method NativeColumnarToRowInfo (line 61) | public NativeColumnarToRowInfo(long memoryAddress, int[] offsets, int[... method numRows (line 72) | public int numRows() { FILE: spark/src/main/java/org/apache/parquet/filter2/predicate/SparkFilterApi.java class SparkFilterApi (line 30) | public final class SparkFilterApi { method intColumn (line 31) | public static IntColumn intColumn(String[] path) { method longColumn (line 35) | public static LongColumn longColumn(String[] path) { method floatColumn (line 39) | public static FloatColumn floatColumn(String[] path) { method doubleColumn (line 43) | public static DoubleColumn doubleColumn(String[] path) { method booleanColumn (line 47) | public static BooleanColumn booleanColumn(String[] path) { method binaryColumn (line 51) | public static BinaryColumn binaryColumn(String[] path) { FILE: spark/src/main/java/org/apache/spark/CometTaskMemoryManager.java class CometTaskMemoryManager (line 36) | public class CometTaskMemoryManager { method CometTaskMemoryManager (line 49) | public CometTaskMemoryManager(long id, long taskAttemptId) { method acquireMemory (line 58) | public long acquireMemory(long size) { method releaseMemory (line 80) | public void releaseMemory(long size) { method getUsed (line 95) | public long getUsed() { class NativeMemoryConsumer (line 104) | private class NativeMemoryConsumer extends MemoryConsumer { method NativeMemoryConsumer (line 105) | protected NativeMemoryConsumer() { method spill (line 109) | @Override method toString (line 115) | @Override FILE: spark/src/main/java/org/apache/spark/shuffle/comet/CometBoundedShuffleMemoryAllocator.java class CometBoundedShuffleMemoryAllocator (line 53) | public final class CometBoundedShuffleMemoryAllocator extends CometShuff... method CometBoundedShuffleMemoryAllocator (line 72) | CometBoundedShuffleMemoryAllocator( method _acquireMemory (line 80) | private synchronized long _acquireMemory(long size) { method spill (line 93) | public long spill(long l, MemoryConsumer memoryConsumer) throws IOExce... method allocateArray (line 97) | public synchronized LongArray allocateArray(long size) { method freeArray (line 103) | public synchronized void freeArray(LongArray array) { method allocate (line 110) | public synchronized MemoryBlock allocate(long required) { method allocateMemoryBlock (line 115) | private synchronized MemoryBlock allocateMemoryBlock(long required) { method free (line 149) | public synchronized long free(MemoryBlock block) { method getOffsetInPage (line 170) | public long getOffsetInPage(long pagePlusOffsetAddress) { method decodeOffset (line 179) | public long decodeOffset(long pagePlusOffsetAddress) { method encodePageNumberAndOffset (line 183) | public long encodePageNumberAndOffset(int pageNumber, long offsetInPag... method encodePageNumberAndOffset (line 188) | public long encodePageNumberAndOffset(MemoryBlock page, long offsetInP... FILE: spark/src/main/java/org/apache/spark/shuffle/comet/CometShuffleChecksumSupport.java type CometShuffleChecksumSupport (line 25) | public interface CometShuffleChecksumSupport { method createPartitionChecksums (line 28) | default long[] createPartitionChecksums(int numPartitions, SparkConf c... method getChecksumAlgorithm (line 43) | default String getChecksumAlgorithm(SparkConf conf) { FILE: spark/src/main/java/org/apache/spark/shuffle/comet/CometShuffleMemoryAllocator.java class CometShuffleMemoryAllocator (line 30) | public final class CometShuffleMemoryAllocator { method getInstance (line 38) | public static CometShuffleMemoryAllocatorTrait getInstance( FILE: spark/src/main/java/org/apache/spark/shuffle/comet/CometShuffleMemoryAllocatorTrait.java class CometShuffleMemoryAllocatorTrait (line 28) | public abstract class CometShuffleMemoryAllocatorTrait extends MemoryCon... method CometShuffleMemoryAllocatorTrait (line 29) | protected CometShuffleMemoryAllocatorTrait( method allocate (line 34) | public abstract MemoryBlock allocate(long required); method free (line 36) | public abstract long free(MemoryBlock block); method getOffsetInPage (line 38) | public abstract long getOffsetInPage(long pagePlusOffsetAddress); method encodePageNumberAndOffset (line 40) | public abstract long encodePageNumberAndOffset(MemoryBlock page, long ... FILE: spark/src/main/java/org/apache/spark/shuffle/comet/CometUnifiedShuffleMemoryAllocator.java class CometUnifiedShuffleMemoryAllocator (line 39) | public final class CometUnifiedShuffleMemoryAllocator extends CometShuff... method CometUnifiedShuffleMemoryAllocator (line 41) | CometUnifiedShuffleMemoryAllocator(TaskMemoryManager taskMemoryManager... method spill (line 51) | public long spill(long l, MemoryConsumer memoryConsumer) throws IOExce... method allocate (line 56) | public synchronized MemoryBlock allocate(long required) { method free (line 60) | public synchronized long free(MemoryBlock block) { method getOffsetInPage (line 75) | public long getOffsetInPage(long pagePlusOffsetAddress) { method encodePageNumberAndOffset (line 79) | public long encodePageNumberAndOffset(int pageNumber, long offsetInPag... method encodePageNumberAndOffset (line 83) | public long encodePageNumberAndOffset(MemoryBlock page, long offsetInP... FILE: spark/src/main/java/org/apache/spark/shuffle/comet/TooLargePageException.java class TooLargePageException (line 22) | public class TooLargePageException extends RuntimeException { method TooLargePageException (line 23) | TooLargePageException(long size) { FILE: spark/src/main/java/org/apache/spark/shuffle/sort/CometShuffleExternalSorter.java type CometShuffleExternalSorter (line 47) | public interface CometShuffleExternalSorter { method getChecksums (line 52) | long[] getChecksums(); method spill (line 55) | void spill() throws IOException; method getPeakMemoryUsedBytes (line 58) | long getPeakMemoryUsedBytes(); method cleanupResources (line 61) | void cleanupResources(); method insertRecord (line 68) | void insertRecord(Object recordBase, long recordOffset, int length, in... method closeAndGetSpills (line 77) | SpillInfo[] closeAndGetSpills() throws IOException; method create (line 85) | static CometShuffleExternalSorter create( FILE: spark/src/main/java/org/apache/spark/shuffle/sort/CometShuffleExternalSorterAsync.java class CometShuffleExternalSorterAsync (line 70) | public final class CometShuffleExternalSorterAsync method CometShuffleExternalSorterAsync (line 126) | public CometShuffleExternalSorterAsync( method createSpillSorter (line 165) | private SpillSorter createSpillSorter() { method getChecksums (line 182) | @Override method spill (line 188) | @Override method getMemoryUsage (line 250) | private long getMemoryUsage() { method updatePeakMemoryUsed (line 261) | private void updatePeakMemoryUsed() { method getPeakMemoryUsedBytes (line 269) | @Override method freeMemory (line 275) | private long freeMemory() { method cleanupResources (line 289) | @Override method growPointerArrayIfNecessary (line 305) | private void growPointerArrayIfNecessary() throws IOException { method insertRecord (line 344) | @Override method closeAndGetSpills (line 376) | @Override FILE: spark/src/main/java/org/apache/spark/shuffle/sort/CometShuffleExternalSorterSync.java class CometShuffleExternalSorterSync (line 61) | public final class CometShuffleExternalSorterSync method CometShuffleExternalSorterSync (line 107) | public CometShuffleExternalSorterSync( method createSpillSorter (line 142) | private SpillSorter createSpillSorter() { method getChecksums (line 159) | @Override method spill (line 165) | @Override method getMemoryUsage (line 200) | private long getMemoryUsage() { method updatePeakMemoryUsed (line 207) | private void updatePeakMemoryUsed() { method getPeakMemoryUsedBytes (line 215) | @Override method freeMemory (line 221) | private long freeMemory() { method cleanupResources (line 229) | @Override method growPointerArrayIfNecessary (line 245) | private void growPointerArrayIfNecessary() throws IOException { method insertRecord (line 284) | @Override method closeAndGetSpills (line 316) | @Override FILE: spark/src/main/java/org/apache/spark/shuffle/sort/SpillSorter.java class SpillSorter (line 46) | public class SpillSorter extends SpillWriter { type SpillCallback (line 49) | @FunctionalInterface method onSpillRequired (line 51) | void onSpillRequired() throws IOException; method SpillSorter (line 92) | public SpillSorter( method freeMemory (line 144) | @Override method getMemoryUsage (line 153) | @Override method spill (line 168) | @Override method freeArray (line 174) | public void freeArray() { method reset (line 185) | public void reset() { method setSpillInfo (line 196) | void setSpillInfo(SpillInfo spillInfo) { method numRecords (line 200) | public int numRecords() { method writeSortedFileNative (line 204) | public void writeSortedFileNative(boolean isLastFile, boolean tracingE... method hasSpaceForAnotherRecord (line 334) | public boolean hasSpaceForAnotherRecord() { method expandPointerArray (line 338) | public void expandPointerArray(LongArray newArray) { method insertRecord (line 343) | public void insertRecord(Object recordBase, long recordOffset, int len... FILE: spark/src/main/java/org/apache/spark/sql/comet/CometScalarSubquery.java class CometScalarSubquery (line 31) | public class CometScalarSubquery { method setSubquery (line 41) | public static synchronized void setSubquery(long planId, ScalarSubquer... method removeSubquery (line 49) | public static synchronized void removeSubquery(long planId, ScalarSubq... method getSubquery (line 60) | private static Object getSubquery(Long planId, Long id) { method isNull (line 69) | public static boolean isNull(long planId, long id) { method getBoolean (line 74) | public static boolean getBoolean(long planId, long id) { method getByte (line 79) | public static byte getByte(long planId, long id) { method getShort (line 84) | public static short getShort(long planId, long id) { method getInt (line 89) | public static int getInt(long planId, long id) { method getLong (line 94) | public static long getLong(long planId, long id) { method getFloat (line 99) | public static float getFloat(long planId, long id) { method getDouble (line 104) | public static double getDouble(long planId, long id) { method getDecimal (line 109) | public static byte[] getDecimal(long planId, long id) { method getString (line 114) | public static String getString(long planId, long id) { method getBinary (line 119) | public static byte[] getBinary(long planId, long id) { FILE: spark/src/main/java/org/apache/spark/sql/comet/execution/shuffle/CometBypassMergeSortShuffleWriter.java class CometBypassMergeSortShuffleWriter (line 74) | final class CometBypassMergeSortShuffleWriter extends ShuffleWrite... method CometBypassMergeSortShuffleWriter (line 123) | CometBypassMergeSortShuffleWriter( method write (line 163) | @Override method getPartitionLengths (line 277) | @Override method writePartitionedData (line 287) | private long[] writePartitionedData(ShuffleMapOutputWriter mapOutputWr... method writePartitionedDataWithChannel (line 330) | private void writePartitionedDataWithChannel(File file, WritableByteCh... method writePartitionedDataWithStream (line 346) | private void writePartitionedDataWithStream(File file, ShufflePartitio... method stop (line 365) | @Override FILE: spark/src/main/java/org/apache/spark/sql/comet/execution/shuffle/CometDiskBlockWriter.java class CometDiskBlockWriter (line 68) | public final class CometDiskBlockWriter { method CometDiskBlockWriter (line 129) | CometDiskBlockWriter( method setChecksumAlgo (line 171) | public void setChecksumAlgo(String checksumAlgo) { method setChecksum (line 175) | public void setChecksum(long checksum) { method getChecksum (line 179) | public long getChecksum() { method doSpill (line 183) | private void doSpill(boolean forceSync) throws IOException { method getOutputRecords (line 241) | public long getOutputRecords() { method insertRow (line 246) | public void insertRow(UnsafeRow row, int partitionId) throws IOExcepti... method close (line 293) | FileSegment close() throws IOException { method getFile (line 327) | File getFile() { method getActiveMemoryUsage (line 332) | long getActiveMemoryUsage() { method freeMemory (line 336) | void freeMemory() { class ArrowIPCWriter (line 343) | class ArrowIPCWriter extends SpillWriter { method ArrowIPCWriter (line 350) | ArrowIPCWriter() { method insertRecord (line 361) | void insertRecord(Object recordBase, long recordOffset, int length) { method numRecords (line 377) | int numRecords() { method doSpilling (line 382) | long doSpilling(boolean isLast) throws IOException { method spill (line 431) | @Override FILE: spark/src/main/java/org/apache/spark/sql/comet/execution/shuffle/CometUnsafeShuffleWriter.java class CometUnsafeShuffleWriter (line 101) | @Private method CometUnsafeShuffleWriter (line 141) | public CometUnsafeShuffleWriter( method openStreamUnchecked (line 178) | private static OutputStream openStreamUnchecked(ShufflePartitionWriter... method updatePeakMemoryUsed (line 186) | private void updatePeakMemoryUsed() { method getPeakMemoryUsedBytes (line 197) | public long getPeakMemoryUsedBytes() { method write (line 203) | @VisibleForTesting method write (line 208) | @Override method open (line 251) | private void open() { method closeAndWriteOutput (line 274) | @VisibleForTesting method insertRecordIntoSorter (line 294) | @VisibleForTesting method forceSorterToSpill (line 312) | @VisibleForTesting method mergeSpills (line 324) | private long[] mergeSpills(SpillInfo[] spills) throws IOException { method mergeSpillsUsingStandardWriter (line 359) | private long[] mergeSpillsUsingStandardWriter(SpillInfo[] spills) thro... method mergeSpillsWithFileStream (line 428) | private void mergeSpillsWithFileStream(SpillInfo[] spills, ShuffleMapO... method mergeSpillsWithTransferTo (line 497) | private void mergeSpillsWithTransferTo(SpillInfo[] spills, ShuffleMapO... method stop (line 555) | @Override method getPartitionLengths (line 582) | @Override class StreamFallbackChannelWrapper (line 587) | private static final class StreamFallbackChannelWrapper implements Wri... method StreamFallbackChannelWrapper (line 590) | StreamFallbackChannelWrapper(OutputStream fallbackStream) { method channel (line 594) | @Override method close (line 599) | @Override FILE: spark/src/main/java/org/apache/spark/sql/comet/execution/shuffle/ExposedByteArrayOutputStream.java class ExposedByteArrayOutputStream (line 25) | public final class ExposedByteArrayOutputStream extends ByteArrayOutputS... method ExposedByteArrayOutputStream (line 26) | ExposedByteArrayOutputStream(int size) { method getBuf (line 30) | public byte[] getBuf() { FILE: spark/src/main/java/org/apache/spark/sql/comet/execution/shuffle/ShuffleThreadPool.java class ShuffleThreadPool (line 28) | public class ShuffleThreadPool { method getThreadPool (line 32) | public static synchronized ExecutorService getThreadPool() { class ThreadPoolQueue (line 56) | final class ThreadPoolQueue extends ArrayBlockingQueue { method ThreadPoolQueue (line 57) | public ThreadPoolQueue(int capacity) { method offer (line 61) | @Override FILE: spark/src/main/java/org/apache/spark/sql/comet/execution/shuffle/SpillInfo.java class SpillInfo (line 27) | public final class SpillInfo { method SpillInfo (line 32) | public SpillInfo(int numPartitions, File file, TempShuffleBlockId bloc... FILE: spark/src/main/java/org/apache/spark/sql/comet/execution/shuffle/SpillWriter.java class SpillWriter (line 48) | public abstract class SpillWriter { method serializeSchema (line 76) | protected byte[][] serializeSchema(StructType schema) { method setChecksumAlgo (line 94) | protected void setChecksumAlgo(String checksumAlgo) { method setChecksum (line 109) | protected void setChecksum(long checksum) { method getChecksum (line 113) | protected long getChecksum() { method spill (line 122) | protected abstract void spill(int required) throws IOException; method acquireNewPageIfNecessary (line 134) | public boolean acquireNewPageIfNecessary(int required) { method initialCurrentPage (line 157) | public void initialCurrentPage(int required) { method doSpilling (line 171) | protected long doSpilling( method freeMemory (line 220) | public long freeMemory() { method getMemoryUsage (line 233) | public long getMemoryUsage() { FILE: spark/src/test/java/org/apache/comet/hadoop/fs/FakeHDFSFileSystem.java class FakeHDFSFileSystem (line 26) | public class FakeHDFSFileSystem extends RawLocalFileSystem { method FakeHDFSFileSystem (line 30) | public FakeHDFSFileSystem() { method getScheme (line 36) | @Override method getUri (line 41) | @Override FILE: spark/src/test/java/org/apache/iceberg/rest/RESTCatalogAdapter.java class RESTCatalogAdapter (line 76) | public class RESTCatalogAdapter implements RESTClient { method setVendedCredentials (line 105) | public void setVendedCredentials(Map credentials) { method RESTCatalogAdapter (line 109) | public RESTCatalogAdapter(Catalog catalog) { type HTTPMethod (line 116) | enum HTTPMethod { type Route (line 123) | enum Route { method Route (line 191) | Route(HTTPMethod method, String pattern) { method Route (line 195) | Route( method matches (line 223) | private boolean matches(HTTPMethod requestMethod, List reque... method variables (line 234) | private Map variables(List requestPath) { method from (line 240) | public static Pair> from(HTTPMethod metho... method requestClass (line 251) | public Class requestClass() { method responseClass (line 255) | public Class responseClass() { method handleOAuthRequest (line 260) | private static OAuthTokenResponse handleOAuthRequest(Object body) { method handleRequest (line 287) | @SuppressWarnings({"MethodLength", "checkstyle:CyclomaticComplexity"}) method doHandleRequest (line 308) | private T doHandleRequest( method commitTransaction (line 513) | private static void commitTransaction(Catalog catalog, CommitTransacti... method execute (line 538) | public T execute( method delete (line 576) | @Override method delete (line 585) | @Override method post (line 595) | @Override method get (line 605) | @Override method head (line 615) | @Override method postForm (line 620) | @Override method close (line 630) | @Override class BadResponseType (line 637) | private static class BadResponseType extends RuntimeException { method BadResponseType (line 638) | private BadResponseType(Class responseType, Object response) { class BadRequestType (line 644) | private static class BadRequestType extends RuntimeException { method BadRequestType (line 645) | private BadRequestType(Class requestType, Object request) { method castRequest (line 650) | public static T castRequest(Class requestType, Object request) { method castResponse (line 658) | public static T castResponse(Class respons... method configureResponseFromException (line 666) | public static void configureResponseFromException( method namespaceFromPathVars (line 675) | private static Namespace namespaceFromPathVars(Map pat... method identFromPathVars (line 679) | private static TableIdentifier identFromPathVars(Map p... FILE: spark/src/test/resources/sql-tests/expressions/string/concat.sql type test_concat_binary (line 58) | CREATE TABLE test_concat_binary USING parquet AS SELECT cast(uuid() as b... FILE: spark/src/test/spark-3.x/org/apache/iceberg/rest/RESTCatalogServlet.java class RESTCatalogServlet (line 54) | public class RESTCatalogServlet extends HttpServlet { method RESTCatalogServlet (line 61) | public RESTCatalogServlet(RESTCatalogAdapter restCatalogAdapter) { method doGet (line 65) | @Override method doHead (line 71) | @Override method doPost (line 77) | @Override method doDelete (line 83) | @Override method execute (line 89) | protected void execute(ServletRequestContext context, HttpServletRespo... method handle (line 123) | protected Consumer handle(HttpServletResponse response) { class ServletRequestContext (line 134) | public static class ServletRequestContext { method ServletRequestContext (line 144) | private ServletRequestContext(ErrorResponse errorResponse) { method ServletRequestContext (line 148) | private ServletRequestContext( method from (line 163) | static ServletRequestContext from(HttpServletRequest request) throws... method method (line 198) | public HTTPMethod method() { method route (line 202) | public Route route() { method path (line 206) | public String path() { method headers (line 210) | public Map headers() { method queryParams (line 214) | public Map queryParams() { method body (line 218) | public Object body() { method error (line 222) | public Optional error() { FILE: spark/src/test/spark-4.0/org/apache/iceberg/rest/RESTCatalogServlet.java class RESTCatalogServlet (line 54) | public class RESTCatalogServlet extends HttpServlet { method RESTCatalogServlet (line 61) | public RESTCatalogServlet(RESTCatalogAdapter restCatalogAdapter) { method doGet (line 65) | @Override method doHead (line 71) | @Override method doPost (line 77) | @Override method doDelete (line 83) | @Override method execute (line 89) | protected void execute(ServletRequestContext context, HttpServletRespo... method handle (line 123) | protected Consumer handle(HttpServletResponse response) { class ServletRequestContext (line 134) | public static class ServletRequestContext { method ServletRequestContext (line 144) | private ServletRequestContext(ErrorResponse errorResponse) { method ServletRequestContext (line 148) | private ServletRequestContext( method from (line 163) | static ServletRequestContext from(HttpServletRequest request) throws... method method (line 198) | public HTTPMethod method() { method route (line 202) | public Route route() { method path (line 206) | public String path() { method headers (line 210) | public Map headers() { method queryParams (line 214) | public Map queryParams() { method body (line 218) | public Object body() { method error (line 222) | public Optional error() {