SYMBOL INDEX (1874 symbols across 206 files) FILE: .travis/check_script_template.py function main (line 21) | def main(argv): FILE: dev_tools/docker/clusters/apache_2.6.0/images/base/scripts/generate_conf_files.py function add_property (line 6) | def add_property(conf, name, value): function write_xml (line 12) | def write_xml(root, fname): function generate_xml_conf_file (line 20) | def generate_xml_conf_file(fname, props): function generate_core_site (line 27) | def generate_core_site(fname): function generate_hdfs_site (line 35) | def generate_hdfs_site(fname): function generate_yarn_site (line 46) | def generate_yarn_site(fname): function generate_mapred_site (line 68) | def generate_mapred_site(fname): function generate_capacity_scheduler (line 86) | def generate_capacity_scheduler(fname): function main (line 100) | def main(argv): FILE: dev_tools/docker/clusters/apache_2.6.0/images/bootstrap/scripts/bootstrap.py function etc_updated (line 15) | def etc_updated(): function boot_node (line 25) | def boot_node(kz, nodename): function main (line 34) | def main(): FILE: dev_tools/docker/scripts/share_etc_hosts.py class App (line 15) | class App(object): method __init__ (line 16) | def __init__(self, compose_group_name): method _get_containers (line 20) | def _get_containers(self, compose_group_name): method _get_hosts (line 26) | def _get_hosts(self): method share_etc_hosts (line 34) | def share_etc_hosts(self): function docker_client (line 44) | def docker_client(): function main (line 78) | def main(argv): FILE: examples/avro/py/avro_base.py class ColorPickBase (line 29) | class ColorPickBase(api.Mapper): method get_user (line 32) | def get_user(self, ctx): method map (line 40) | def map(self, ctx): class AvroKeyColorPick (line 47) | class AvroKeyColorPick(ColorPickBase): method get_user (line 49) | def get_user(self, ctx): class AvroValueColorPick (line 53) | class AvroValueColorPick(ColorPickBase): method get_user (line 55) | def get_user(self, ctx): class AvroKeyValueColorPick (line 59) | class AvroKeyValueColorPick(ColorPickBase): method get_user (line 61) | def get_user(self, ctx): method map (line 64) | def map(self, ctx): class ColorCountBase (line 69) | class ColorCountBase(api.Reducer): method reduce (line 71) | def reduce(self, ctx): method emit (line 76) | def emit(self, s, ctx): class NoAvroColorCount (line 83) | class NoAvroColorCount(ColorCountBase): method emit (line 85) | def emit(self, s, ctx): class AvroKeyColorCount (line 89) | class AvroKeyColorCount(ColorCountBase): method emit (line 91) | def emit(self, s, ctx): class AvroValueColorCount (line 95) | class AvroValueColorCount(ColorCountBase): method emit (line 97) | def emit(self, s, ctx): class AvroKeyValueColorCount (line 101) | class AvroKeyValueColorCount(ColorCountBase): method emit (line 103) | def emit(self, s, ctx): function run_task (line 108) | def run_task(mapper_class, reducer_class=NoAvroColorCount): FILE: examples/avro/py/avro_container_dump_results.py function main (line 25) | def main(fn, out_fn, avro_mode=''): FILE: examples/avro/py/avro_key_in.py function __main__ (line 24) | def __main__(): FILE: examples/avro/py/avro_key_in_out.py function __main__ (line 24) | def __main__(): FILE: examples/avro/py/avro_key_value_in.py function __main__ (line 24) | def __main__(): FILE: examples/avro/py/avro_key_value_in_out.py function __main__ (line 24) | def __main__(): FILE: examples/avro/py/avro_parquet_dump_results.py class Mapper (line 25) | class Mapper(api.Mapper): method map (line 27) | def map(self, ctx): function __main__ (line 32) | def __main__(): FILE: examples/avro/py/avro_pyrw.py class UserReader (line 32) | class UserReader(AvroReader): class ColorWriter (line 36) | class ColorWriter(AvroWriter): method emit (line 40) | def emit(self, key, value): class ColorPick (line 44) | class ColorPick(api.Mapper): method map (line 46) | def map(self, ctx): class ColorCount (line 53) | class ColorCount(api.Reducer): method reduce (line 55) | def reduce(self, ctx): FILE: examples/avro/py/avro_value_in.py function __main__ (line 24) | def __main__(): FILE: examples/avro/py/avro_value_in_out.py function __main__ (line 24) | def __main__(): FILE: examples/avro/py/check_cc.py function iter_fnames (line 29) | def iter_fnames(path): function main (line 40) | def main(in_, out_): FILE: examples/avro/py/check_results.py function iter_lines (line 27) | def iter_lines(path): function main (line 39) | def main(exp, res): FILE: examples/avro/py/color_count.py class Mapper (line 28) | class Mapper(api.Mapper): method map (line 30) | def map(self, ctx): class Reducer (line 37) | class Reducer(api.Reducer): method reduce (line 39) | def reduce(self, ctx): function __main__ (line 44) | def __main__(): FILE: examples/avro/py/create_input.py function create_input (line 27) | def create_input(n, stream): function main (line 36) | def main(n, filename): FILE: examples/avro/py/gen_data.py class Mapper (line 27) | class Mapper(api.Mapper): method map (line 29) | def map(self, ctx): function __main__ (line 35) | def __main__(): FILE: examples/avro/py/generate_avro_users.py function main (line 37) | def main(argv): FILE: examples/avro/py/kmer_count.py function window (line 28) | def window(s, width): class Mapper (line 33) | class Mapper(api.Mapper): method map (line 35) | def map(self, ctx): class Reducer (line 41) | class Reducer(api.Reducer): method reduce (line 43) | def reduce(self, ctx): function __main__ (line 47) | def __main__(): FILE: examples/avro/py/show_kmer_count.py function main (line 26) | def main(argv): FILE: examples/avro/py/write_avro.py function main (line 30) | def main(schema_fn, csv_fn, avro_fn): FILE: examples/avro/src/main/java/it/crs4/pydoop/WriteKV.java class WriteKV (line 41) | class WriteKV { method buildUser (line 45) | private static GenericRecord buildUser( method buildPet (line 54) | private static GenericRecord buildPet( method createFile (line 62) | private static File createFile(File file, Schema schema, T... reco... method createInputFile (line 74) | private static File createInputFile( method main (line 106) | public static void main(String[] args) throws Exception { FILE: examples/avro/src/main/java/it/crs4/pydoop/WriteParquet.java class WriteParquet (line 53) | public class WriteParquet extends Configured implements Tool { method getSchema (line 60) | private static Schema getSchema(Configuration conf) class WriteUserMap (line 70) | public static class WriteUserMap method setup (line 75) | @Override method map (line 81) | @Override method run (line 94) | public int run(String[] args) throws Exception { method main (line 129) | public static void main(String[] args) throws Exception { FILE: examples/c++/HadoopPipes.cc type HadoopPipes (line 48) | namespace HadoopPipes { class JobConfImpl (line 50) | class JobConfImpl: public JobConf { method set (line 54) | void set(const string& key, const string& value) { method hasKey (line 58) | virtual bool hasKey(const string& key) const { method string (line 62) | virtual const string& get(const string& key) const { method getInt (line 70) | virtual int getInt(const string& key) const { method getFloat (line 75) | virtual float getFloat(const string& key) const { method getBoolean (line 80) | virtual bool getBoolean(const string&key) const { class DownwardProtocol (line 86) | class DownwardProtocol { class UpwardProtocol (line 101) | class UpwardProtocol { class Protocol (line 116) | class Protocol { class TextUpwardProtocol (line 123) | class TextUpwardProtocol: public UpwardProtocol { method writeBuffer (line 129) | void writeBuffer(const string& buffer) { method TextUpwardProtocol (line 134) | TextUpwardProtocol(FILE* _stream): stream(_stream) {} method output (line 136) | virtual void output(const string& key, const string& value) { method partitionedOutput (line 144) | virtual void partitionedOutput(int reduce, const string& key, method status (line 154) | virtual void status(const string& message) { method progress (line 159) | virtual void progress(float progress) { method registerCounter (line 164) | virtual void registerCounter(int id, const string& group, method incrementCounter (line 171) | virtual void incrementCounter(const TaskContext::Counter* counter, method done (line 177) | virtual void done() { class TextProtocol (line 182) | class TextProtocol: public Protocol { method readUpto (line 190) | int readUpto(string& buffer, const char* limit) { method TextProtocol (line 205) | TextProtocol(FILE* down, DownwardProtocol* _handler, FILE* up) { method UpwardProtocol (line 211) | UpwardProtocol* getUplink() { method nextEvent (line 215) | virtual void nextEvent() { type MESSAGE_TYPE (line 297) | enum MESSAGE_TYPE {START_MESSAGE, SET_JOB_CONF, SET_INPUT_TYPES, RUN_MAP, class BinaryUpwardProtocol (line 303) | class BinaryUpwardProtocol: public UpwardProtocol { method BinaryUpwardProtocol (line 307) | BinaryUpwardProtocol(FILE* _stream) { method authenticate (line 312) | virtual void authenticate(const string &responseDigest) { method output (line 318) | virtual void output(const string& key, const string& value) { method partitionedOutput (line 324) | virtual void partitionedOutput(int reduce, const string& key, method status (line 332) | virtual void status(const string& message) { method progress (line 337) | virtual void progress(float progress) { method done (line 343) | virtual void done() { method registerCounter (line 347) | virtual void registerCounter(int id, const string& group, method incrementCounter (line 355) | virtual void incrementCounter(const TaskContext::Counter* counter, class BinaryProtocol (line 367) | class BinaryProtocol: public Protocol { method getPassword (line 376) | void getPassword(string &password) { method verifyDigestAndRespond (line 396) | void verifyDigestAndRespond(string& digest, string& challenge) { method verifyDigest (line 413) | bool verifyDigest(string &password, string& digest, string& challeng... method string (line 422) | string createDigest(string &password, string& msg) { method BinaryProtocol (line 462) | BinaryProtocol(FILE* down, DownwardProtocol* _handler, FILE* up) { method UpwardProtocol (line 471) | UpwardProtocol* getUplink() { method nextEvent (line 475) | virtual void nextEvent() { class CombineContext (line 575) | class CombineContext: public ReduceContext { method CombineContext (line 589) | CombineContext(ReduceContext* _baseContext, method JobConf (line 604) | virtual const JobConf* getJobConf() { method emit (line 616) | virtual void emit(const std::string& key, const std::string& value) { method progress (line 625) | virtual void progress() { method setStatus (line 629) | virtual void setStatus(const std::string& status) { method nextKey (line 633) | bool nextKey() { method nextValue (line 648) | virtual bool nextValue() { method Counter (line 657) | virtual Counter* getCounter(const std::string& group, method incrementCounter (line 662) | virtual void incrementCounter(const Counter* counter, uint64_t amoun... class CombineRunner (line 671) | class CombineRunner: public RecordWriter { method CombineRunner (line 682) | CombineRunner(int64_t _spillSize, ReduceContext* _baseContext, method emit (line 694) | virtual void emit(const std::string& key, method close (line 703) | virtual void close() { method spillAll (line 708) | void spillAll() { class TaskContextImpl (line 719) | class TaskContextImpl: public MapContext, public ReduceContext, method TaskContextImpl (line 751) | TaskContextImpl(const Factory& _factory) { method setProtocol (line 774) | void setProtocol(Protocol* _protocol, UpwardProtocol* _uplink) { method start (line 780) | virtual void start(int protocol) { method setJobConf (line 787) | virtual void setJobConf(vector values) { method setInputTypes (line 797) | virtual void setInputTypes(string keyType, string valueType) { method runMap (line 802) | virtual void runMap(string _inputSplit, int _numReduces, bool pipedI... method mapItem (line 828) | virtual void mapItem(const string& _key, const string& _value) { method runReduce (line 834) | virtual void runReduce(int reduce, bool pipedOutput) { method reduceKey (line 843) | virtual void reduceKey(const string& _key) { method reduceValue (line 848) | virtual void reduceValue(const string& _value) { method isDone (line 853) | virtual bool isDone() { method close (line 860) | virtual void close() { method abort (line 866) | virtual void abort() { method waitForTask (line 870) | void waitForTask() { method nextKey (line 876) | bool nextKey() { method nextValue (line 906) | virtual bool nextValue() { method JobConf (line 919) | virtual JobConf* getJobConf() { method string (line 927) | virtual const string& getInputKey() { method string (line 936) | virtual const string& getInputValue() { method progress (line 944) | virtual void progress() { method setStatus (line 961) | virtual void setStatus(const string& status) { method string (line 970) | virtual const string& getInputKeyClass() { method string (line 977) | virtual const string& getInputValueClass() { method emit (line 988) | virtual void emit(const string& key, const string& value) { method Counter (line 1003) | virtual Counter* getCounter(const std::string& group, method incrementCounter (line 1014) | virtual void incrementCounter(const Counter* counter, uint64_t amoun... method closeAll (line 1018) | void closeAll() { function runTask (line 1104) | bool runTask(const Factory& factory) { FILE: examples/c++/SerialUtils.cc type HadoopUtils (line 29) | namespace HadoopUtils { function serializeInt (line 178) | void serializeInt(int32_t t, OutStream& stream) { function serializeLong (line 182) | void serializeLong(int64_t t, OutStream& stream) function deserializeInt (line 213) | int32_t deserializeInt(InStream& stream) { function deserializeLong (line 217) | int64_t deserializeLong(InStream& stream) function serializeFloat (line 246) | void serializeFloat(float t, OutStream& stream) function deserializeFloat (line 255) | float deserializeFloat(InStream& stream) function deserializeFloat (line 262) | void deserializeFloat(float& t, InStream& stream) function serializeString (line 271) | void serializeString(const std::string& t, OutStream& stream) function deserializeString (line 279) | void deserializeString(std::string& t, InStream& stream) FILE: examples/c++/StringUtils.cc type HadoopUtils (line 32) | namespace HadoopUtils { function string (line 34) | string toString(int32_t x) { function toInt (line 40) | int toInt(const string& val) { function toFloat (line 49) | float toFloat(const string& val) { function toBool (line 58) | bool toBool(const string& val) { function getCurrentMillis (line 72) | uint64_t getCurrentMillis() { function splitString (line 80) | vector splitString(const std::string& str, function string (line 97) | string quoteString(const string& str, function string (line 129) | string unquoteString(const string& str) { FILE: examples/c++/include/hadoop/Pipes.hh class ReduceContext (line 141) | class ReduceContext: public TaskContext { class Closable (line 149) | class Closable { method close (line 151) | virtual void close() {} class Mapper (line 158) | class Mapper: public Closable { class Reducer (line 166) | class Reducer: public Closable { class Partitioner (line 174) | class Partitioner { class RecordReader (line 184) | class RecordReader: public Closable { class RecordWriter (line 198) | class RecordWriter: public Closable { class Factory (line 207) | class Factory { method Reducer (line 216) | virtual Reducer* createCombiner(MapContext& context) const { method Partitioner (line 225) | virtual Partitioner* createPartitioner(MapContext& context) const { method RecordReader (line 234) | virtual RecordReader* createRecordReader(MapContext& context) const { method RecordWriter (line 243) | virtual RecordWriter* createRecordWriter(ReduceContext& context) const { FILE: examples/c++/include/hadoop/SerialUtils.hh type HadoopUtils (line 24) | namespace HadoopUtils { class Error (line 29) | class Error { class InStream (line 68) | class InStream { class OutStream (line 83) | class OutStream { class FileInStream (line 102) | class FileInStream : public InStream { class FileOutStream (line 125) | class FileOutStream: public OutStream { class StringInStream (line 151) | class StringInStream: public InStream { FILE: examples/c++/include/hadoop/StringUtils.hh type HadoopUtils (line 25) | namespace HadoopUtils { FILE: examples/c++/include/hadoop/TemplateFactory.hh type HadoopPipes (line 21) | namespace HadoopPipes { class TemplateFactory2 (line 24) | class TemplateFactory2: public Factory { method Mapper (line 26) | Mapper* createMapper(MapContext& context) const { method Reducer (line 29) | Reducer* createReducer(ReduceContext& context) const { class TemplateFactory3 (line 35) | class TemplateFactory3: public TemplateFactory2 { method Partitioner (line 37) | Partitioner* createPartitioner(MapContext& context) const { class TemplateFactory3 (line 43) | class TemplateFactory3 class TemplateFactory4 (line 48) | class TemplateFactory4 method Reducer (line 51) | Reducer* createCombiner(MapContext& context) const { class TemplateFactory4 (line 57) | class TemplateFactory4 class TemplateFactory5 (line 63) | class TemplateFactory5 method RecordReader (line 66) | RecordReader* createRecordReader(MapContext& context) const { class TemplateFactory5 (line 72) | class TemplateFactory5 class TemplateFactory (line 79) | class TemplateFactory method RecordWriter (line 82) | RecordWriter* createRecordWriter(ReduceContext& context) const { class TemplateFactory (line 89) | class TemplateFactory method configure (line 34) | public void configure(JobConf conf) { method isSplitable (line 38) | protected boolean isSplitable(FileSystem fs, Path file) { method getRecordReader (line 42) | public RecordReader getRecordReader( FILE: examples/input_format/it/crs4/pydoop/mapreduce/TextInputFormat.java class TextInputFormat (line 33) | public class TextInputFormat extends FileInputFormat { method createRecordReader (line 35) | @Override method isSplitable (line 41) | @Override FILE: examples/pydoop_script/check.py function check_base_histogram (line 43) | def check_base_histogram(mr_out_dir): function check_caseswitch (line 58) | def check_caseswitch(mr_out_dir, switch="upper"): function check_grep (line 67) | def check_grep(mr_out_dir): function check_lowercase (line 79) | def check_lowercase(mr_out_dir): function check_transpose (line 83) | def check_transpose(mr_out_dir): function check_wordcount (line 104) | def check_wordcount(mr_out_dir, stop_words=None): function check_wordcount_sw (line 111) | def check_wordcount_sw(mr_out_dir): function make_parser (line 120) | def make_parser(): function main (line 128) | def main(argv): FILE: examples/pydoop_script/scripts/base_histogram.py function mapper (line 27) | def mapper(_, samrecord, writer): function reducer (line 34) | def reducer(key, ivalue, writer): FILE: examples/pydoop_script/scripts/caseswitch.py function mapper (line 28) | def mapper(_, record, writer, conf): FILE: examples/pydoop_script/scripts/grep.py function mapper (line 31) | def mapper(_, text, writer, conf): FILE: examples/pydoop_script/scripts/lowercase.py function mapper (line 27) | def mapper(_, record, writer): FILE: examples/pydoop_script/scripts/transpose.py function mapper (line 60) | def mapper(key, value, writer): function reducer (line 70) | def reducer(key, ivalue, writer): FILE: examples/pydoop_script/scripts/wc_combiner.py function mapper (line 24) | def mapper(_, text, writer): function reducer (line 29) | def reducer(word, icounts, writer): function combiner (line 34) | def combiner(word, icounts, writer): FILE: examples/pydoop_script/scripts/wordcount.py function mapper (line 25) | def mapper(_, text, writer): function reducer (line 30) | def reducer(word, icounts, writer): FILE: examples/pydoop_script/scripts/wordcount_sw.py function mapper (line 33) | def mapper(_, value, writer): function reducer (line 41) | def reducer(word, icounts, writer): FILE: examples/pydoop_submit/check.py function check_wordcount_minimal (line 41) | def check_wordcount_minimal(mr_out_dir): function check_nosep (line 51) | def check_nosep(mr_out_dir): function check_map_only_python_writer (line 65) | def check_map_only_python_writer(mr_out_dir): FILE: examples/pydoop_submit/mr/map_only_java_writer.py class Mapper (line 27) | class Mapper(api.Mapper): method __init__ (line 29) | def __init__(self, context): method map (line 32) | def map(self, context): function __main__ (line 36) | def __main__(): FILE: examples/pydoop_submit/mr/map_only_python_writer.py class Mapper (line 32) | class Mapper(api.Mapper): method __init__ (line 34) | def __init__(self, context): method map (line 37) | def map(self, context): class Writer (line 41) | class Writer(api.RecordWriter): method __init__ (line 43) | def __init__(self, context): method close (line 53) | def close(self): method emit (line 57) | def emit(self, key, value): function __main__ (line 61) | def __main__(): FILE: examples/pydoop_submit/mr/nosep.py class Mapper (line 25) | class Mapper(api.Mapper): method map (line 27) | def map(self, ctx): function __main__ (line 32) | def __main__(): FILE: examples/pydoop_submit/mr/wordcount_full.py class Mapper (line 34) | class Mapper(api.Mapper): method __init__ (line 36) | def __init__(self, context): method map (line 41) | def map(self, context): class Reducer (line 48) | class Reducer(api.Reducer): method __init__ (line 50) | def __init__(self, context): method reduce (line 55) | def reduce(self, context): class Reader (line 60) | class Reader(api.RecordReader): method __init__ (line 65) | def __init__(self, context): method close (line 81) | def close(self): method next (line 86) | def next(self): method get_progress (line 96) | def get_progress(self): class Writer (line 100) | class Writer(api.RecordWriter): method __init__ (line 102) | def __init__(self, context): method close (line 112) | def close(self): method emit (line 117) | def emit(self, key, value): class Partitioner (line 121) | class Partitioner(api.Partitioner): method __init__ (line 123) | def __init__(self, context): method partition (line 127) | def partition(self, key, n_reduces): function main (line 145) | def main(): FILE: examples/pydoop_submit/mr/wordcount_minimal.py class Mapper (line 30) | class Mapper(api.Mapper): method map (line 32) | def map(self, context): class Reducer (line 37) | class Reducer(api.Reducer): method reduce (line 39) | def reduce(self, context): function main (line 46) | def main(): FILE: examples/self_contained/check_results.py function compute_vc (line 31) | def compute_vc(input_dir): function get_res (line 41) | def get_res(output_dir): function check (line 45) | def check(measured_res, expected_res): function main (line 53) | def main(argv): FILE: examples/self_contained/vowelcount/lib/__init__.py function is_vowel (line 23) | def is_vowel(c): FILE: examples/self_contained/vowelcount/mr/main.py function main (line 24) | def main(): FILE: examples/self_contained/vowelcount/mr/mapper.py class Mapper (line 23) | class Mapper(api.Mapper): method map (line 25) | def map(self, context): FILE: examples/self_contained/vowelcount/mr/reducer.py class Reducer (line 22) | class Reducer(api.Reducer): method reduce (line 24) | def reduce(self, context): FILE: examples/sequence_file/bin/filter.py class FilterMapper (line 31) | class FilterMapper(Mapper): method __init__ (line 36) | def __init__(self, context): method map (line 41) | def map(self, context): function __main__ (line 48) | def __main__(): FILE: examples/sequence_file/bin/wordcount.py class WordCountMapper (line 27) | class WordCountMapper(Mapper): method map (line 29) | def map(self, context): class WordCountReducer (line 34) | class WordCountReducer(Reducer): method reduce (line 36) | def reduce(self, context): function __main__ (line 41) | def __main__(): FILE: examples/sequence_file/check.py function main (line 27) | def main(args): FILE: int_test/mapred_submitter/check.py function get_lines (line 30) | def get_lines(dir_path): function check_output (line 41) | def check_output(items, exp_items): function check_counters (line 53) | def check_counters(counter, exp_counter): function word_count (line 57) | def word_count(lines): function check_map_only (line 61) | def check_map_only(in_dir, out_dir): function check_map_reduce (line 67) | def check_map_reduce(in_dir, out_dir): function check_pstats (line 74) | def check_pstats(pstats_dir): FILE: int_test/mapred_submitter/genwords.py function genfile (line 38) | def genfile(path, size): FILE: int_test/mapred_submitter/mr/map_only_java_writer.py class Mapper (line 25) | class Mapper(api.Mapper): method map (line 27) | def map(self, context): function __main__ (line 31) | def __main__(): FILE: int_test/mapred_submitter/mr/map_only_python_writer.py class Mapper (line 28) | class Mapper(api.Mapper): method map (line 30) | def map(self, context): class Writer (line 34) | class Writer(api.RecordWriter): method __init__ (line 36) | def __init__(self, context): method close (line 42) | def close(self): method emit (line 45) | def emit(self, key, value): function __main__ (line 49) | def __main__(): FILE: int_test/mapred_submitter/mr/map_reduce_combiner.py class Mapper (line 25) | class Mapper(api.Mapper): method map (line 27) | def map(self, context): class Reducer (line 32) | class Reducer(api.Reducer): method reduce (line 34) | def reduce(self, context): function __main__ (line 38) | def __main__(): FILE: int_test/mapred_submitter/mr/map_reduce_java_rw.py class Mapper (line 25) | class Mapper(api.Mapper): method map (line 27) | def map(self, context): class Reducer (line 32) | class Reducer(api.Reducer): method reduce (line 34) | def reduce(self, context): function __main__ (line 38) | def __main__(): FILE: int_test/mapred_submitter/mr/map_reduce_java_rw_pstats.py class Mapper (line 25) | class Mapper(api.Mapper): method map (line 27) | def map(self, context): class Reducer (line 32) | class Reducer(api.Reducer): method reduce (line 34) | def reduce(self, context): function __main__ (line 38) | def __main__(): FILE: int_test/mapred_submitter/mr/map_reduce_python_partitioner.py class Mapper (line 27) | class Mapper(api.Mapper): method map (line 29) | def map(self, context): class Reducer (line 34) | class Reducer(api.Reducer): method reduce (line 36) | def reduce(self, context): class Partitioner (line 40) | class Partitioner(api.Partitioner): method partition (line 42) | def partition(self, key, n_reduces): function __main__ (line 46) | def __main__(): FILE: int_test/mapred_submitter/mr/map_reduce_python_reader.py class Mapper (line 28) | class Mapper(api.Mapper): method map (line 30) | def map(self, context): class Reducer (line 35) | class Reducer(api.Reducer): method reduce (line 37) | def reduce(self, context): class Reader (line 41) | class Reader(api.RecordReader): method __init__ (line 43) | def __init__(self, context): method close (line 53) | def close(self): method next (line 56) | def next(self): method get_progress (line 66) | def get_progress(self): function __main__ (line 70) | def __main__(): FILE: int_test/mapred_submitter/mr/map_reduce_python_writer.py class Mapper (line 28) | class Mapper(api.Mapper): method map (line 30) | def map(self, context): class Reducer (line 35) | class Reducer(api.Reducer): method reduce (line 37) | def reduce(self, context): class Writer (line 41) | class Writer(api.RecordWriter): method __init__ (line 43) | def __init__(self, context): method close (line 49) | def close(self): method emit (line 52) | def emit(self, key, value): function __main__ (line 56) | def __main__(): FILE: int_test/mapred_submitter/mr/map_reduce_raw_io.py class Mapper (line 25) | class Mapper(api.Mapper): method map (line 29) | def map(self, context): class Reducer (line 36) | class Reducer(api.Reducer): method reduce (line 38) | def reduce(self, context): function __main__ (line 43) | def __main__(): FILE: int_test/mapred_submitter/mr/map_reduce_slow_java_rw.py class Mapper (line 28) | class Mapper(api.Mapper): method __init__ (line 30) | def __init__(self, context): method map (line 34) | def map(self, context): method close (line 40) | def close(self): class Reducer (line 44) | class Reducer(api.Reducer): method __init__ (line 46) | def __init__(self, context): method reduce (line 50) | def reduce(self, context): method close (line 55) | def close(self): function __main__ (line 59) | def __main__(): FILE: int_test/mapred_submitter/mr/map_reduce_slow_python_rw.py class Mapper (line 33) | class Mapper(api.Mapper): method __init__ (line 35) | def __init__(self, context): method map (line 39) | def map(self, context): method close (line 45) | def close(self): class Reducer (line 49) | class Reducer(api.Reducer): method __init__ (line 51) | def __init__(self, context): method reduce (line 55) | def reduce(self, context): method close (line 60) | def close(self): class Reader (line 64) | class Reader(api.RecordReader): method __init__ (line 66) | def __init__(self, context): method close (line 76) | def close(self): method next (line 79) | def next(self): method get_progress (line 89) | def get_progress(self): class Writer (line 93) | class Writer(api.RecordWriter): method __init__ (line 95) | def __init__(self, context): method close (line 101) | def close(self): method emit (line 104) | def emit(self, key, value): function __main__ (line 108) | def __main__(): FILE: int_test/opaque_split/check.py function check_output (line 29) | def check_output(mr_out_dir): FILE: int_test/opaque_split/gen_splits.py function gen_ranges (line 32) | def gen_ranges(): FILE: int_test/opaque_split/mrapp.py class Reader (line 44) | class Reader(api.RecordReader): method __init__ (line 46) | def __init__(self, context): method next (line 53) | def next(self): method get_progress (line 57) | def get_progress(self): class Mapper (line 62) | class Mapper(api.Mapper): method map (line 64) | def map(self, context): function __main__ (line 68) | def __main__(): FILE: int_test/progress/mrapp.py class Mapper (line 28) | class Mapper(api.Mapper): method map (line 30) | def map(self, context): class Writer (line 36) | class Writer(api.RecordWriter): method __init__ (line 38) | def __init__(self, context): method close (line 46) | def close(self): method emit (line 49) | def emit(self, key, value): function __main__ (line 56) | def __main__(): FILE: pydoop/__init__.py function reset (line 62) | def reset(): function hadoop_home (line 66) | def hadoop_home(): function hadoop_conf (line 70) | def hadoop_conf(): function hadoop_params (line 74) | def hadoop_params(): function hadoop_classpath (line 78) | def hadoop_classpath(): function package_dir (line 82) | def package_dir(): function jar_name (line 92) | def jar_name(hadoop_vinfo=None): function jar_path (line 96) | def jar_path(hadoop_vinfo=None): function complete_mod_name (line 104) | def complete_mod_name(module, hadoop_vinfo=None): function import_version_specific_module (line 108) | def import_version_specific_module(name): class AddSectionWrapper (line 119) | class AddSectionWrapper(object): method __init__ (line 123) | def __init__(self, f): method __iter__ (line 127) | def __iter__(self): method __next__ (line 130) | def __next__(self): method readline (line 136) | def readline(self): function read_properties (line 146) | def read_properties(fname): class LocalModeNotSupported (line 159) | class LocalModeNotSupported(RuntimeError): method __init__ (line 160) | def __init__(self): function check_local_mode (line 165) | def check_local_mode(): FILE: pydoop/app/argparse_types.py function kv_pair (line 23) | def kv_pair(s): class UpdateMap (line 31) | class UpdateMap(argparse.Action): method __init__ (line 42) | def __init__(self, option_strings, dest, **kwargs): method __call__ (line 47) | def __call__(self, parser, namespace, values, option_string=None): function a_file_that_can_be_read (line 53) | def a_file_that_can_be_read(x): function a_hdfs_file (line 59) | def a_hdfs_file(x): function a_comma_separated_list (line 64) | def a_comma_separated_list(x): FILE: pydoop/app/main.py class PatchedArgumentParser (line 39) | class PatchedArgumentParser(argparse.ArgumentParser): method _read_args_from_files (line 45) | def _read_args_from_files(self, arg_strings): function make_parser (line 54) | def make_parser(): function main (line 72) | def main(argv=None): FILE: pydoop/app/script.py class PydoopScript (line 45) | class PydoopScript(object): method __init__ (line 47) | def __init__(self, args, unknown_args): method generate_driver (line 53) | def generate_driver(mr_module, args): method convert_args (line 64) | def convert_args(self, args, unknown_args): method run (line 112) | def run(self): method clean (line 122) | def clean(self): function run (line 126) | def run(args, unknown_args=None): function add_parser_arguments (line 135) | def add_parser_arguments(parser): function add_parser (line 151) | def add_parser(subparsers): FILE: pydoop/app/submit.py class PydoopSubmitter (line 54) | class PydoopSubmitter(object): method __init__ (line 60) | def __init__(self): method __cache_archive_link (line 80) | def __cache_archive_link(archive_name): method __set_files_to_cache_helper (line 84) | def __set_files_to_cache_helper(self, prop, upload_and_cache, cache): method __set_files_to_cache (line 103) | def __set_files_to_cache(self, args): method __set_archives_to_cache (line 110) | def __set_archives_to_cache(self, args): method _env_arg_to_dict (line 120) | def _env_arg_to_dict(set_env_list): method set_args (line 132) | def set_args(self, args, unknown_args=None): method __warn_user_if_wd_maybe_unreadable (line 163) | def __warn_user_if_wd_maybe_unreadable(self, abs_remote_path): method _generate_pipes_code (line 201) | def _generate_pipes_code(self): method __validate (line 272) | def __validate(self): method __clean_wd (line 282) | def __clean_wd(self): method __setup_remote_paths (line 292) | def __setup_remote_paths(self): method run (line 324) | def run(self): method fake_run_class (line 384) | def fake_run_class(self, *args, **kwargs): function run (line 391) | def run(args, unknown_args=None): function add_parser_common_arguments (line 400) | def add_parser_common_arguments(parser): function add_parser_arguments (line 481) | def add_parser_arguments(parser): function add_parser (line 550) | def add_parser(subparsers): FILE: pydoop/avrolib.py class Deserializer (line 39) | class Deserializer(object): method __init__ (line 41) | def __init__(self, schema_str): method deserialize (line 45) | def deserialize(self, rec_bytes): class Serializer (line 49) | class Serializer(object): method __init__ (line 51) | def __init__(self, schema_str): method serialize (line 55) | def serialize(self, record): class SeekableDataFileReader (line 73) | class SeekableDataFileReader(DataFileReader): method align_after (line 77) | def align_after(self, offset): class AvroReader (line 102) | class AvroReader(RecordReader): method __init__ (line 108) | def __init__(self, ctx): method next (line 117) | def next(self): method get_progress (line 124) | def get_progress(self): class AvroWriter (line 135) | class AvroWriter(RecordWriter): method __init__ (line 139) | def __init__(self, context): method close (line 148) | def close(self): FILE: pydoop/hadoop_utils.py class HadoopXMLError (line 35) | class HadoopXMLError(Exception): function extract_text (line 39) | def extract_text(node): function parse_hadoop_conf_file (line 45) | def parse_hadoop_conf_file(fn): class PathFinder (line 67) | class PathFinder(object): method __init__ (line 71) | def __init__(self): method reset (line 78) | def reset(self): method hadoop_home (line 82) | def hadoop_home(self): method hadoop_conf (line 95) | def hadoop_conf(self): method hadoop_params (line 110) | def hadoop_params(self): method hadoop_classpath (line 123) | def hadoop_classpath(self): method __get_is_local (line 134) | def __get_is_local(self): method is_local (line 144) | def is_local(self): FILE: pydoop/hadut.py function _pop_generic_args (line 43) | def _pop_generic_args(args): function _merge_csv_args (line 59) | def _merge_csv_args(args): function _construct_property_args (line 79) | def _construct_property_args(prop_dict): class RunCmdError (line 84) | class RunCmdError(RuntimeError): method __init__ (line 89) | def __init__(self, returncode, cmd, output=None): method __str__ (line 94) | def __str__(self): function run_tool_cmd (line 105) | def run_tool_cmd(tool, cmd, args=None, properties=None, hadoop_conf_dir=... function run_cmd (line 155) | def run_cmd(cmd, args=None, properties=None, hadoop_home=None, function run_class (line 167) | def run_class(class_name, args=None, properties=None, classpath=None, function iter_mr_out_files (line 209) | def iter_mr_out_files(mr_out_dir): function collect_output (line 215) | def collect_output(mr_out_dir, out_file=None): FILE: pydoop/hdfs/__init__.py function init (line 83) | def init(): function reset (line 92) | def reset(): function open (line 101) | def open(hdfs_path, mode="r", buff_size=0, replication=0, blocksize=0, function dump (line 116) | def dump(data, hdfs_path, **kwargs): function load (line 133) | def load(hdfs_path, **kwargs): function _cp_file (line 149) | def _cp_file(src_fs, src_path, dest_fs, dest_path, **kwargs): function cp (line 164) | def cp(src_hdfs_path, dest_hdfs_path, **kwargs): function put (line 220) | def put(src_path, dest_hdfs_path, **kwargs): function get (line 232) | def get(src_hdfs_path, dest_path, **kwargs): function mkdir (line 244) | def mkdir(hdfs_path, user=None): function rm (line 255) | def rm(hdfs_path, recursive=True, user=None): function rmr (line 270) | def rmr(hdfs_path, user=None): function lsl (line 274) | def lsl(hdfs_path, user=None, recursive=False): function ls (line 300) | def ls(hdfs_path, user=None, recursive=False): function chmod (line 311) | def chmod(hdfs_path, mode, user=None): function move (line 327) | def move(src, dest, user=None): function chown (line 343) | def chown(hdfs_path, user=None, group=None, hdfs_user=None): function rename (line 354) | def rename(from_path, to_path, user=None): function renames (line 368) | def renames(from_path, to_path, user=None): FILE: pydoop/hdfs/common.py function parse_mode (line 45) | def parse_mode(mode): function encode_path (line 60) | def encode_path(path): function decode_path (line 63) | def decode_path(path): function encode_host (line 66) | def encode_host(host): function decode_host (line 69) | def decode_host(host): function encode_path (line 72) | def encode_path(path): function decode_path (line 77) | def decode_path(path): function encode_host (line 82) | def encode_host(host): function decode_host (line 87) | def decode_host(host): function get_groups (line 93) | def get_groups(user=DEFAULT_USER): FILE: pydoop/hdfs/core/__init__.py function init (line 26) | def init(): function core_hdfs_fs (line 38) | def core_hdfs_fs(host, port, user): FILE: pydoop/hdfs/file.py function _complain_ifclosed (line 31) | def _complain_ifclosed(closed): class FileIO (line 36) | class FileIO(object): method __init__ (line 47) | def __init__(self, raw_hdfs_file, fs, mode, encoding=None, errors=None): method __enter__ (line 76) | def __enter__(self): method __exit__ (line 79) | def __exit__(self, exc_type, exc_value, traceback): method fs (line 83) | def fs(self): method name (line 90) | def name(self): method size (line 97) | def size(self): method writable (line 104) | def writable(self): method readline (line 107) | def readline(self): method next (line 122) | def next(self): method __next__ (line 129) | def __next__(self): method __iter__ (line 140) | def __iter__(self): method available (line 143) | def available(self): method close (line 154) | def close(self): method pread (line 165) | def pread(self, position, length): method read (line 188) | def read(self, length=-1): method seek (line 218) | def seek(self, position, whence=os.SEEK_SET): method tell (line 232) | def tell(self): method write (line 242) | def write(self, data): method flush (line 258) | def flush(self): class hdfs_file (line 266) | class hdfs_file(FileIO): method pread_chunk (line 268) | def pread_chunk(self, position, chunk): method read_chunk (line 286) | def read_chunk(self, chunk): class local_file (line 301) | class local_file(io.FileIO): method __init__ (line 309) | def __init__(self, fs, name, mode): method __make_parents (line 320) | def __make_parents(fs, name): method fs (line 329) | def fs(self): method size (line 333) | def size(self): method available (line 336) | def available(self): method close (line 340) | def close(self): method seek (line 347) | def seek(self, position, whence=os.SEEK_SET): method __seek_and_read (line 352) | def __seek_and_read(self, position, length=None, buf=None): method pread (line 366) | def pread(self, position, length): method pread_chunk (line 369) | def pread_chunk(self, position, chunk): method read_chunk (line 372) | def read_chunk(self, chunk): class TextIOWrapper (line 377) | class TextIOWrapper(io.TextIOWrapper): method __getattr__ (line 379) | def __getattr__(self, name): method pread (line 390) | def pread(self, position, length): FILE: pydoop/hdfs/fs.py class _FSStatus (line 45) | class _FSStatus(object): method __init__ (line 47) | def __init__(self, fs, host, port, user, refcount=1): method __repr__ (line 54) | def __repr__(self): function _complain_ifclosed (line 58) | def _complain_ifclosed(closed): function _get_ip (line 63) | def _get_ip(host, default=None): function _get_connection_info (line 71) | def _get_connection_info(host, port, user): function _default_fs (line 92) | def _default_fs(): function default_is_local (line 98) | def default_is_local(): class hdfs (line 108) | class hdfs(object): method __canonize_hpu (line 134) | def __canonize_hpu(self, hpu): method __lookup (line 141) | def __lookup(self, hpu): method __eq__ (line 146) | def __eq__(self, other): method __init__ (line 153) | def __init__(self, host="default", port=0, user=None, groups=None): method __enter__ (line 187) | def __enter__(self): method __exit__ (line 190) | def __exit__(self, exc_type, exc_value, traceback): method fs (line 194) | def fs(self): method refcount (line 198) | def refcount(self): method host (line 202) | def host(self): method port (line 209) | def port(self): method user (line 216) | def user(self): method close (line 222) | def close(self): method closed (line 234) | def closed(self): method open_file (line 237) | def open_file(self, path, method capacity (line 284) | def capacity(self): method copy (line 296) | def copy(self, from_path, to_hdfs, to_path): method create_directory (line 313) | def create_directory(self, path): method default_block_size (line 325) | def default_block_size(self): method delete (line 335) | def delete(self, path, recursive=True): method exists (line 350) | def exists(self, path): method get_hosts (line 362) | def get_hosts(self, path, start, length): method get_path_info (line 380) | def get_path_info(self, path): method list_directory (line 407) | def list_directory(self, path): method move (line 420) | def move(self, from_path, to_hdfs, to_path): method rename (line 437) | def rename(self, from_path, to_path): method set_replication (line 450) | def set_replication(self, path, replication): method set_working_directory (line 463) | def set_working_directory(self, path): method used (line 475) | def used(self): method working_directory (line 485) | def working_directory(self): method chown (line 496) | def chown(self, path, user='', group=''): method __get_umask (line 512) | def __get_umask(): method __compute_mode_from_string (line 517) | def __compute_mode_from_string(self, path, mode_string): method chmod (line 580) | def chmod(self, path, mode): method utime (line 597) | def utime(self, path, mtime, atime): method walk (line 612) | def walk(self, top): FILE: pydoop/hdfs/path.py class StatResult (line 35) | class StatResult(object): method __init__ (line 49) | def __init__(self, path_info): method __repr__ (line 72) | def __repr__(self): class _HdfsPathSplitter (line 81) | class _HdfsPathSplitter(object): method raise_bad_path (line 86) | def raise_bad_path(cls, hdfs_path, why=None): method parse (line 92) | def parse(cls, hdfs_path): method unparse (line 118) | def unparse(cls, scheme, netloc, path): method split_netloc (line 132) | def split_netloc(cls, netloc): method split (line 150) | def split(cls, hdfs_path, user): function parse (line 173) | def parse(hdfs_path): function unparse (line 185) | def unparse(scheme, netloc, path): function split (line 192) | def split(hdfs_path, user=None): function join (line 208) | def join(*parts): function abspath (line 242) | def abspath(hdfs_path, user=None, local=False): function splitpath (line 281) | def splitpath(hdfs_path): function basename (line 289) | def basename(hdfs_path): function dirname (line 296) | def dirname(hdfs_path): function exists (line 304) | def exists(hdfs_path, user=None): function lstat (line 316) | def lstat(hdfs_path, user=None): function lexists (line 320) | def lexists(hdfs_path, user=None): function kind (line 325) | def kind(path, user=None): function isdir (line 341) | def isdir(path, user=None): function isfile (line 348) | def isfile(path, user=None): function expanduser (line 355) | def expanduser(path): function expandvars (line 371) | def expandvars(path): function _update_stat (line 378) | def _update_stat(st, path_): function stat (line 389) | def stat(path, user=None): function getatime (line 403) | def getatime(path, user=None): function getmtime (line 410) | def getmtime(path, user=None): function getctime (line 417) | def getctime(path, user=None): function getsize (line 424) | def getsize(path, user=None): function isfull (line 431) | def isfull(path): function isabs (line 441) | def isabs(path): function islink (line 452) | def islink(path, user=None): function ismount (line 464) | def ismount(path): function normcase (line 476) | def normcase(path): function normpath (line 480) | def normpath(path): function realpath (line 488) | def realpath(path): function samefile (line 500) | def samefile(path1, path2, user=None): function splitdrive (line 509) | def splitdrive(path): function splitext (line 513) | def splitext(path): function access (line 520) | def access(path, mode, user=None): function utime (line 544) | def utime(hdfs_path, times=None, user=None): FILE: pydoop/jc.py function jc_wrapper (line 24) | def jc_wrapper(obj): FILE: pydoop/mapreduce/api.py class JobConf (line 35) | class JobConf(dict): method get_int (line 48) | def get_int(self, key, default=None): method get_float (line 55) | def get_float(self, key, default=None): method get_bool (line 62) | def get_bool(self, key, default=None): method get_json (line 83) | def get_json(self, key, default=None): class InputSplit (line 88) | class InputSplit(object): class FileSplit (line 98) | class FileSplit(InputSplit, class OpaqueSplit (line 106) | class OpaqueSplit(InputSplit, namedtuple("OpaqueSplit", "payload")): class Context (line 124) | class Context(ABC): method input_split (line 139) | def input_split(self): method get_input_split (line 152) | def get_input_split(self, raw=False): method job_conf (line 156) | def job_conf(self): method get_job_conf (line 163) | def get_job_conf(self): method key (line 167) | def key(self): method get_input_key (line 174) | def get_input_key(self): method value (line 178) | def value(self): method get_input_value (line 185) | def get_input_value(self): method values (line 189) | def values(self): method get_input_values (line 196) | def get_input_values(self): method emit (line 200) | def emit(self, key, value): method progress (line 207) | def progress(self): method set_status (line 211) | def set_status(self, status): method get_counter (line 221) | def get_counter(self, group, name): method increment_counter (line 235) | def increment_counter(self, counter, amount): class Closable (line 242) | class Closable(object): method close (line 244) | def close(self): class Component (line 253) | class Component(ABC): method __init__ (line 255) | def __init__(self, context): class Mapper (line 259) | class Mapper(Component, Closable): method map (line 265) | def map(self, context): class Reducer (line 279) | class Reducer(Component, Closable): method reduce (line 286) | def reduce(self, context): class Combiner (line 299) | class Combiner(Reducer): class Partitioner (line 315) | class Partitioner(Component): method partition (line 326) | def partition(self, key, num_of_reduces): class RecordReader (line 342) | class RecordReader(Component, Closable): method __iter__ (line 347) | def __iter__(self): method next (line 351) | def next(self): method __next__ (line 364) | def __next__(self): method get_progress (line 368) | def get_progress(self): class RecordWriter (line 379) | class RecordWriter(Component, Closable): method emit (line 385) | def emit(self, key, value): class Factory (line 397) | class Factory(ABC): method create_mapper (line 414) | def create_mapper(self, context): method create_reducer (line 417) | def create_reducer(self, context): method create_combiner (line 420) | def create_combiner(self, context): method create_partitioner (line 428) | def create_partitioner(self, context): method create_record_reader (line 437) | def create_record_reader(self, context): method create_record_writer (line 446) | def create_record_writer(self, context): FILE: pydoop/mapreduce/binary_protocol.py function get_password (line 85) | def get_password(): function _get_LongWritable (line 97) | def _get_LongWritable(downlink): function _get_Text (line 102) | def _get_Text(downlink): function _get_avro_key (line 112) | def _get_avro_key(downlink): function _get_avro_value (line 117) | def _get_avro_value(downlink): function _get_pickled (line 122) | def _get_pickled(downlink): class Downlink (line 126) | class Downlink(object): method __init__ (line 183) | def __init__(self, istream, context, **kwargs): method close (line 193) | def close(self): method read_job_conf (line 196) | def read_job_conf(self): method verify_digest (line 203) | def verify_digest(self, digest, challenge): method setup_record_writer (line 209) | def setup_record_writer(self, piped_output): method get_k (line 216) | def get_k(self): method get_v (line 219) | def get_v(self): method setup_avro_deser (line 222) | def setup_avro_deser(self): method setup_deser (line 240) | def setup_deser(self, key_type, value_type): method __next__ (line 250) | def __next__(self): method __iter__ (line 343) | def __iter__(self): method next (line 347) | def next(self): class Uplink (line 351) | class Uplink(object): method __init__ (line 356) | def __init__(self, stream): method flush (line 359) | def flush(self): method close (line 362) | def close(self): method authenticate (line 367) | def authenticate(self, response_digest): method output (line 370) | def output(self, k, v): method partitioned_output (line 373) | def partitioned_output(self, part, k, v): method status (line 376) | def status(self, msg): method progress (line 379) | def progress(self, p): method done (line 382) | def done(self): method register_counter (line 385) | def register_counter(self, id, group, name): method increment_counter (line 388) | def increment_counter(self, id, amount): FILE: pydoop/mapreduce/connections.py class Connection (line 38) | class Connection(object): method __init__ (line 56) | def __init__(self, context, istream, ostream, **kwargs): method close (line 60) | def close(self): method __enter__ (line 64) | def __enter__(self): method __exit__ (line 67) | def __exit__(self, *args): class NetworkConnection (line 71) | class NetworkConnection(Connection): method __init__ (line 73) | def __init__(self, context, host, port, **kwargs): method close (line 82) | def close(self): class FileConnection (line 87) | class FileConnection(Connection): method __init__ (line 89) | def __init__(self, context, in_fn, out_fn, **kwargs): function get_connection (line 97) | def get_connection(context, **kwargs): FILE: pydoop/mapreduce/pipes.py function create_digest (line 60) | def create_digest(key, msg): function read_int_writable (line 67) | def read_int_writable(f): function write_int_writable (line 72) | def write_int_writable(n, f): function read_bytes_writable (line 76) | def read_bytes_writable(f): function write_bytes_writable (line 84) | def write_bytes_writable(s, f): class FileSplit (line 90) | class FileSplit(api.FileSplit): method frombuffer (line 93) | def frombuffer(cls, buf): class OpaqueSplit (line 98) | class OpaqueSplit(api.OpaqueSplit): method frombuffer (line 101) | def frombuffer(cls, buf): method read (line 105) | def read(cls, f): method write (line 108) | def write(self, f): function write_opaque_splits (line 112) | def write_opaque_splits(splits, f): function read_opaque_splits (line 118) | def read_opaque_splits(f): class TaskContext (line 123) | class TaskContext(api.Context): method __init__ (line 129) | def __init__(self, factory, **kwargs): method get_input_split (line 159) | def get_input_split(self, raw=False): method get_job_conf (line 169) | def get_job_conf(self): method get_input_key (line 172) | def get_input_key(self): method get_input_value (line 175) | def get_input_value(self): method get_input_values (line 178) | def get_input_values(self): method create_combiner (line 181) | def create_combiner(self): method create_mapper (line 190) | def create_mapper(self): method create_partitioner (line 194) | def create_partitioner(self): method create_record_reader (line 198) | def create_record_reader(self): method create_record_writer (line 202) | def create_record_writer(self): method create_reducer (line 206) | def create_reducer(self): method progress (line 210) | def progress(self): method set_status (line 227) | def set_status(self, status): method get_counter (line 231) | def get_counter(self, group, name): method increment_counter (line 238) | def increment_counter(self, counter, amount): method __spill_counters (line 244) | def __spill_counters(self): method _authenticate (line 250) | def _authenticate(self, password, digest, challenge): method _setup_avro_ser (line 257) | def _setup_avro_ser(self): method __maybe_serialize (line 273) | def __maybe_serialize(self, key, value): method emit (line 286) | def emit(self, key, value): method __actual_emit (line 309) | def __actual_emit(self, key, value): method __spill_all (line 320) | def __spill_all(self): method close (line 330) | def close(self): method get_output_dir (line 352) | def get_output_dir(self): method get_work_path (line 355) | def get_work_path(self): method get_task_partition (line 361) | def get_task_partition(self): method get_default_work_file (line 364) | def get_default_work_file(self, extension=""): class Factory (line 374) | class Factory(api.Factory): method __init__ (line 376) | def __init__(self, mapper_class, method create_mapper (line 389) | def create_mapper(self, context): method create_reducer (line 392) | def create_reducer(self, context): method create_combiner (line 395) | def create_combiner(self, context): method create_partitioner (line 398) | def create_partitioner(self, context): method create_record_reader (line 401) | def create_record_reader(self, context): method create_record_writer (line 404) | def create_record_writer(self, context): function _run (line 408) | def _run(context, **kwargs): function run_task (line 414) | def run_task(factory, **kwargs): FILE: pydoop/test_support.py function __inject_pos (line 33) | def __inject_pos(code, start=0): function inject_code (line 45) | def inject_code(new_code, target_code): function add_sys_path (line 59) | def add_sys_path(target_code): function set_python_cmd (line 67) | def set_python_cmd(code, python_cmd=sys.executable): function adapt_script (line 77) | def adapt_script(code, python_cmd=sys.executable): function parse_mr_output (line 81) | def parse_mr_output(output, vtype=str): function compare_counts (line 97) | def compare_counts(c1, c2): class LocalWordCount (line 111) | class LocalWordCount(object): method __init__ (line 113) | def __init__(self, input_path, min_occurrence=0, stop_words=None): method expected_output (line 120) | def expected_output(self): method run (line 125) | def run(self): method _wordcount_file (line 138) | def _wordcount_file(self, wc, fn, path=None): method check (line 145) | def check(self, output): function get_wd_prefix (line 155) | def get_wd_prefix(base="pydoop_"): FILE: pydoop/test_utils.py function _get_special_chr (line 49) | def _get_special_chr(): class FSTree (line 83) | class FSTree(object): method __init__ (line 99) | def __init__(self, name, kind=1): method add (line 106) | def add(self, name, kind=1): method walk (line 111) | def walk(self): function make_wd (line 119) | def make_wd(fs, prefix="pydoop_test_"): function make_random_data (line 128) | def make_random_data(size=_RANDOM_DATA_SIZE, printable=True): function get_bytes_per_checksum (line 134) | def get_bytes_per_checksum(): function silent_call (line 141) | def silent_call(func, *args, **kwargs): function get_module (line 155) | def get_module(name, path=None): function compile_java (line 165) | def compile_java(java_file, classpath, opts=None): function run_java (line 184) | def run_java(jclass, classpath, args, wd): function get_java_output_stream (line 192) | def get_java_output_stream(jclass, classpath, args, wd): class WDTestCase (line 199) | class WDTestCase(unittest.TestCase): method setUp (line 201) | def setUp(self): method tearDown (line 204) | def tearDown(self): method _mkfn (line 207) | def _mkfn(self, basename): method _mkf (line 210) | def _mkf(self, basename, mode='w'): FILE: pydoop/utils/jvm.py function get_java_home (line 37) | def get_java_home(): function load_jvm_lib (line 74) | def load_jvm_lib(java_home=None): function get_include_dirs (line 85) | def get_include_dirs(): function get_libraries (line 101) | def get_libraries(): function get_macros (line 114) | def get_macros(): function get_jvm_lib_path_and_name (line 125) | def get_jvm_lib_path_and_name(java_home=None): function check_jni_header (line 139) | def check_jni_header(include_dirs=None): function find_file (line 150) | def find_file(path, to_find): FILE: pydoop/utils/misc.py class NullHandler (line 31) | class NullHandler(logging.Handler): method emit (line 32) | def emit(self, record): class NullLogger (line 36) | class NullLogger(logging.Logger): method __init__ (line 37) | def __init__(self): function make_random_str (line 43) | def make_random_str(prefix="pydoop_", postfix=''): class Timer (line 47) | class Timer(object): method __init__ (line 49) | def __init__(self, ctx, counter_group=None): method _gen_counter_name (line 55) | def _gen_counter_name(self, event): method _get_time_counter (line 58) | def _get_time_counter(self, name): method start (line 66) | def start(self, s): method stop (line 69) | def stop(self, s): method time_block (line 73) | def time_block(self, event_name): class TimingBlock (line 76) | class TimingBlock(object): method __init__ (line 78) | def __init__(self, timer, event_name): method __enter__ (line 82) | def __enter__(self): method __exit__ (line 86) | def __exit__(self, exception_type, exception_val, exception_tb): FILE: pydoop/utils/py3compat.py class Py2ABC (line 43) | class Py2ABC(object): function __identity (line 47) | def __identity(x): function __chr (line 51) | def __chr(x): function __iteritems_2 (line 55) | def __iteritems_2(x): function __iteritems_3 (line 59) | def __iteritems_3(x): function __parser_read_2 (line 63) | def __parser_read_2(parser, f): function __parser_read_3 (line 67) | def __parser_read_3(parser, f): FILE: setup.py function rm_rf (line 90) | def rm_rf(path, dry_run=False): function mtime (line 108) | def mtime(fn): function must_generate (line 112) | def must_generate(target, prerequisites): function get_version_string (line 119) | def get_version_string(): function write_config (line 127) | def write_config(filename="pydoop/config.py"): function write_version (line 137) | def write_version(filename="pydoop/version.py"): class JavaLib (line 179) | class JavaLib(object): method __init__ (line 181) | def __init__(self): class JavaBuilder (line 194) | class JavaBuilder(object): method __init__ (line 196) | def __init__(self, build_temp, build_lib): method run (line 201) | def run(self): method __build_java_lib (line 205) | def __build_java_lib(self, jlib): class BuildPydoopExt (line 251) | class BuildPydoopExt(build_ext): method __have_better_tls (line 253) | def __have_better_tls(self): method __finalize_hdfs (line 269) | def __finalize_hdfs(self, ext): method build_extension (line 292) | def build_extension(self, ext): class BuildPydoop (line 298) | class BuildPydoop(build): method build_java (line 300) | def build_java(self): method create_tmp (line 304) | def create_tmp(self): method clean_up (line 310) | def clean_up(self): method run (line 313) | def run(self): FILE: src/it/crs4/pydoop/NoSeparatorTextOutputFormat.java class NoSeparatorTextOutputFormat (line 41) | public class NoSeparatorTextOutputFormat extends TextOutputFormat FILE: src/it/crs4/pydoop/mapreduce/pipes/Application.java class Application (line 74) | class Application context, method getSecurityChallenge (line 163) | private String getSecurityChallenge() { method writePasswordToLocalFile (line 174) | private void writePasswordToLocalFile(String localPasswordFile, method getDownlink (line 190) | DownwardProtocol getDownlink() { method waitForAuthentication (line 199) | void waitForAuthentication() throws IOException, method waitForFinish (line 211) | boolean waitForFinish() throws Throwable { method abort (line 221) | void abort(Throwable t) throws IOException { method cleanup (line 243) | void cleanup() throws IOException { method runClient (line 260) | static Process runClient(List command, method createDigest (line 270) | public static String createDigest(byte[] password, String data) FILE: src/it/crs4/pydoop/mapreduce/pipes/BinaryProtocol.java class BinaryProtocol (line 51) | class BinaryProtocol { method authenticate (line 40) | void authenticate(String digest, String challenge) throws IOException; method start (line 46) | void start() throws IOException; method setJobConf (line 53) | void setJobConf(Configuration conf) throws IOException; method setInputTypes (line 61) | void setInputTypes(String keyType, String valueType) throws IOException; method runMap (line 70) | void runMap(InputSplit split, int numReduces, method mapItem (line 79) | void mapItem(K key, V value) throws IOException; method runReduce (line 87) | void runReduce(int reduce, boolean pipedOutput) throws IOException; method reduceKey (line 94) | void reduceKey(K key) throws IOException; method reduceValue (line 101) | void reduceValue(V value) throws IOException; method endOfInput (line 108) | void endOfInput() throws IOException; method abort (line 114) | void abort() throws IOException; method flush (line 119) | void flush() throws IOException; method close (line 124) | void close() throws IOException, InterruptedException; FILE: src/it/crs4/pydoop/mapreduce/pipes/DummyRecordReader.java class DummyRecordReader (line 27) | public abstract class DummyRecordReader method next (line 30) | public abstract boolean next(FloatWritable key, NullWritable value) FILE: src/it/crs4/pydoop/mapreduce/pipes/OpaqueSplit.java class OpaqueSplit (line 33) | class OpaqueSplit extends InputSplit implements Writable { method OpaqueSplit (line 37) | public OpaqueSplit() { method OpaqueSplit (line 41) | public OpaqueSplit(byte[] payload) { method getPayload (line 45) | public BytesWritable getPayload() { method getLength (line 49) | @Override method toString (line 54) | @Override method getLocations (line 59) | @Override method getLocationInfo (line 64) | @Override method write (line 71) | @Override method readFields (line 76) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/OutputHandler.java class OutputHandler (line 36) | class OutputHandler method OutputHandler (line 57) | public OutputHandler(TaskInputOutputContext context, method output (line 68) | @Override method partitionedOutput (line 76) | @Override method status (line 86) | @Override method progress (line 96) | @Override method done (line 109) | @Override method getProgress (line 121) | public float getProgress() { method failed (line 128) | public void failed(Throwable e) { method waitForFinish (line 140) | public synchronized boolean waitForFinish() throws Throwable { method registerCounter (line 150) | @Override method incrementCounter (line 156) | @Override method authenticate (line 166) | public synchronized boolean authenticate(String digest) throws IOExcep... method waitForAuthentication (line 184) | synchronized void waitForAuthentication() FILE: src/it/crs4/pydoop/mapreduce/pipes/PipesMapper.java class PipesMapper (line 40) | class PipesMapper getSplits(JobContext context) method getOpaqueSplits (line 69) | private List getOpaqueSplits(Configuration conf, String uri) method createRecordReader (line 92) | @Override class PipesDummyRecordReader (line 109) | static class PipesDummyRecordReader extends DummyRecordReader { method PipesDummyRecordReader (line 113) | public PipesDummyRecordReader() {} method PipesDummyRecordReader (line 115) | public PipesDummyRecordReader(InputSplit split, TaskAttemptContext c... method initialize (line 120) | @Override method close (line 124) | public synchronized void close() throws IOException {} method getProgress (line 126) | @Override method nextKeyValue (line 131) | @Override method getCurrentKey (line 136) | @Override method getCurrentValue (line 142) | @Override method next (line 148) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PipesNonJavaOutputFormat.java class PipesNonJavaOutputFormat (line 29) | public class PipesNonJavaOutputFormat extends FileOutputFormat method setConf (line 42) | public void setConf(Configuration conf) { method getConf (line 48) | public Configuration getConf() { method setNextPartition (line 56) | static void setNextPartition(int newValue) { method getPartition (line 67) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PipesReducer.java class PipesReducer (line 38) | class PipesReducer getInRecords() method initialize (line 54) | public void initialize(InputSplit split, TaskAttemptContext context) method getCurrentKey (line 64) | @Override method getCurrentValue (line 71) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroBridgeKeyValueReader.java class PydoopAvroBridgeKeyValueReader (line 36) | public class PydoopAvroBridgeKeyValueReader method PydoopAvroBridgeKeyValueReader (line 41) | public PydoopAvroBridgeKeyValueReader( method getInRecords (line 48) | protected List getInRecords() method initialize (line 55) | public void initialize(InputSplit split, TaskAttemptContext context) method getCurrentKey (line 67) | @Override method getCurrentValue (line 74) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroBridgeKeyValueWriter.java class PydoopAvroBridgeKeyValueWriter (line 33) | public class PydoopAvroBridgeKeyValueWriter method PydoopAvroBridgeKeyValueWriter (line 36) | public PydoopAvroBridgeKeyValueWriter( method write (line 43) | public void write(Text key, Text value) FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroBridgeKeyWriter.java class PydoopAvroBridgeKeyWriter (line 34) | public class PydoopAvroBridgeKeyWriter extends PydoopAvroBridgeWriterBase { method PydoopAvroBridgeKeyWriter (line 36) | public PydoopAvroBridgeKeyWriter( method write (line 43) | public void write(Text key, Text ignore) FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroBridgeReaderBase.java class PydoopAvroBridgeReaderBase (line 42) | public abstract class PydoopAvroBridgeReaderBase method getInRecords (line 71) | protected abstract List getInRecords() method initialize (line 74) | public void initialize(InputSplit split, TaskAttemptContext context) method nextKeyValue (line 105) | public synchronized boolean nextKeyValue() method getProgress (line 147) | public float getProgress() throws IOException, InterruptedException { method close (line 151) | public synchronized void close() throws IOException { FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroBridgeValueReader.java class PydoopAvroBridgeValueReader (line 37) | public class PydoopAvroBridgeValueReader method PydoopAvroBridgeValueReader (line 42) | public PydoopAvroBridgeValueReader( method getInRecords (line 48) | protected List getInRecords() method initialize (line 54) | public void initialize(InputSplit split, TaskAttemptContext context) method getCurrentKey (line 64) | @Override method getCurrentValue (line 70) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroBridgeValueWriter.java class PydoopAvroBridgeValueWriter (line 34) | public class PydoopAvroBridgeValueWriter extends PydoopAvroBridgeWriterB... method PydoopAvroBridgeValueWriter (line 36) | public PydoopAvroBridgeValueWriter( method write (line 43) | public void write(Text ignore, Text value) FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroBridgeWriterBase.java class PydoopAvroBridgeWriterBase (line 46) | public abstract class PydoopAvroBridgeWriterBase method PydoopAvroBridgeWriterBase (line 64) | public PydoopAvroBridgeWriterBase(TaskAttemptContext context, AvroIO m... method getOutRecords (line 91) | protected List getOutRecords(List inRecords) method write (line 104) | protected void write(List outRecords) method close (line 126) | public void close(TaskAttemptContext context) FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroInputBridgeBase.java class PydoopAvroInputBridgeBase (line 32) | public abstract class PydoopAvroInputBridgeBase method getActualFormat (line 38) | protected InputFormat getActualFormat(Configuration conf) { method getSplits (line 49) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroInputKeyBridge.java class PydoopAvroInputKeyBridge (line 31) | public class PydoopAvroInputKeyBridge method PydoopAvroInputKeyBridge (line 34) | public PydoopAvroInputKeyBridge() { method createRecordReader (line 38) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroInputKeyValueBridge.java class PydoopAvroInputKeyValueBridge (line 30) | public class PydoopAvroInputKeyValueBridge method PydoopAvroInputKeyValueBridge (line 33) | public PydoopAvroInputKeyValueBridge() { method createRecordReader (line 37) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroInputValueBridge.java class PydoopAvroInputValueBridge (line 31) | public class PydoopAvroInputValueBridge method PydoopAvroInputValueBridge (line 34) | public PydoopAvroInputValueBridge() { method createRecordReader (line 38) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroKeyInputFormat.java class PydoopAvroKeyInputFormat (line 32) | public class PydoopAvroKeyInputFormat method createRecordReader (line 35) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroKeyOutputFormat.java class PydoopAvroKeyOutputFormat (line 30) | public class PydoopAvroKeyOutputFormat method getRecordWriter (line 33) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroKeyRecordReader.java class PydoopAvroKeyRecordReader (line 31) | public class PydoopAvroKeyRecordReader method PydoopAvroKeyRecordReader (line 37) | public PydoopAvroKeyRecordReader(Schema readerSchema) { method getCurrentKey (line 41) | @Override method getCurrentValue (line 47) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroKeyRecordWriter.java class PydoopAvroKeyRecordWriter (line 31) | public class PydoopAvroKeyRecordWriter method PydoopAvroKeyRecordWriter (line 34) | public PydoopAvroKeyRecordWriter(Schema writerSchema, method write (line 40) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroKeyValueInputFormat.java class PydoopAvroKeyValueInputFormat (line 31) | public class PydoopAvroKeyValueInputFormat method createRecordReader (line 34) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroKeyValueOutputFormat.java class PydoopAvroKeyValueOutputFormat (line 31) | public class PydoopAvroKeyValueOutputFormat method getRecordWriter (line 34) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroKeyValueRecordReader.java class PydoopAvroKeyValueRecordReader (line 27) | public class PydoopAvroKeyValueRecordReader method PydoopAvroKeyValueRecordReader (line 30) | public PydoopAvroKeyValueRecordReader(Schema readerSchema) { method getCurrentKey (line 34) | @Override method getCurrentValue (line 40) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroKeyValueRecordWriter.java class PydoopAvroKeyValueRecordWriter (line 31) | public class PydoopAvroKeyValueRecordWriter method PydoopAvroKeyValueRecordWriter (line 36) | public PydoopAvroKeyValueRecordWriter(Schema writerSchema, method write (line 43) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroOutputBridgeBase.java class PydoopAvroOutputBridgeBase (line 33) | public abstract class PydoopAvroOutputBridgeBase method getActualFormat (line 39) | protected OutputFormat getActualFormat(Configuration conf) { method checkOutputSpecs (line 50) | @Override method getOutputCommitter (line 57) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroOutputFormatBase.java class PydoopAvroOutputFormatBase (line 31) | public abstract class PydoopAvroOutputFormatBase method getOutputSchema (line 34) | protected static Schema getOutputSchema( FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroOutputKeyBridge.java class PydoopAvroOutputKeyBridge (line 30) | public class PydoopAvroOutputKeyBridge extends PydoopAvroOutputBridgeBase { method PydoopAvroOutputKeyBridge (line 32) | public PydoopAvroOutputKeyBridge() { method getRecordWriter (line 36) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroOutputKeyValueBridge.java class PydoopAvroOutputKeyValueBridge (line 30) | public class PydoopAvroOutputKeyValueBridge method PydoopAvroOutputKeyValueBridge (line 33) | public PydoopAvroOutputKeyValueBridge() { method getRecordWriter (line 37) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroOutputValueBridge.java class PydoopAvroOutputValueBridge (line 30) | public class PydoopAvroOutputValueBridge extends PydoopAvroOutputBridgeB... method PydoopAvroOutputValueBridge (line 32) | public PydoopAvroOutputValueBridge() { method getRecordWriter (line 36) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroRecordReaderBase.java class PydoopAvroRecordReaderBase (line 41) | public abstract class PydoopAvroRecordReaderBase method PydoopAvroRecordReaderBase (line 53) | protected PydoopAvroRecordReaderBase(Schema readerSchema) { method initialize (line 58) | @Override method nextKeyValue (line 77) | @Override method getProgress (line 87) | @Override method close (line 100) | @Override method getCurrentRecord (line 111) | protected GenericRecord getCurrentRecord() { method createSeekableInput (line 115) | protected SeekableInput createSeekableInput(Configuration conf, Path p... FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroRecordWriterBase.java class PydoopAvroRecordWriterBase (line 34) | public abstract class PydoopAvroRecordWriterBase method PydoopAvroRecordWriterBase (line 39) | protected PydoopAvroRecordWriterBase(Schema writerSchema, method close (line 48) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroValueInputFormat.java class PydoopAvroValueInputFormat (line 32) | public class PydoopAvroValueInputFormat method createRecordReader (line 35) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroValueOutputFormat.java class PydoopAvroValueOutputFormat (line 30) | public class PydoopAvroValueOutputFormat method getRecordWriter (line 33) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroValueRecordReader.java class PydoopAvroValueRecordReader (line 31) | public class PydoopAvroValueRecordReader method PydoopAvroValueRecordReader (line 37) | public PydoopAvroValueRecordReader(Schema readerSchema) { method getCurrentKey (line 41) | @Override method getCurrentValue (line 47) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/PydoopAvroValueRecordWriter.java class PydoopAvroValueRecordWriter (line 31) | public class PydoopAvroValueRecordWriter method PydoopAvroValueRecordWriter (line 34) | public PydoopAvroValueRecordWriter(Schema writerSchema, method write (line 40) | @Override FILE: src/it/crs4/pydoop/mapreduce/pipes/Submitter.java class CommandLineParser (line 66) | class CommandLineParser { method CommandLineParser (line 69) | CommandLineParser() { method addOption (line 85) | void addOption(String longName, boolean required, String description, method addArgument (line 93) | void addArgument(String name, boolean required, String description) { method parse (line 100) | CommandLine parse(Configuration conf, String[] args) method printUsage (line 108) | void printUsage() { class Submitter (line 131) | public class Submitter extends Configured implements Tool { type AvroIO (line 133) | public static enum AvroIO { method getPydoopProperties (line 162) | public static Properties getPydoopProperties() { method Submitter (line 176) | public Submitter() { method isLocalFS (line 181) | public static boolean isLocalFS(Configuration conf) throws IOException { method getExecutable (line 190) | public static String getExecutable(Configuration conf) { method setExecutable (line 200) | public static void setExecutable(Configuration conf, String executable) { method setIsJavaRecordReader (line 209) | public static void setIsJavaRecordReader(Configuration conf, boolean v... method getIsJavaRecordReader (line 218) | public static boolean getIsJavaRecordReader(Configuration conf) { method setIsJavaMapper (line 227) | public static void setIsJavaMapper(Configuration conf, boolean value) { method getIsJavaMapper (line 236) | public static boolean getIsJavaMapper(Configuration conf) { method setIsJavaReducer (line 245) | public static void setIsJavaReducer(Configuration conf, boolean value) { method getIsJavaReducer (line 254) | public static boolean getIsJavaReducer(Configuration conf) { method setIsJavaRecordWriter (line 263) | public static void setIsJavaRecordWriter(Configuration conf, boolean v... method getIsJavaRecordWriter (line 272) | public static boolean getIsJavaRecordWriter(Configuration conf) { method setIfUnset (line 283) | private static void setIfUnset(Configuration conf, String key, String ... method setJavaPartitioner (line 294) | static void setJavaPartitioner(Configuration conf, Class cls) { method getJavaPartitioner (line 303) | static Class getJavaPartitioner(Configuration c... method getClass (line 308) | private static method getKeepCommandFile (line 327) | public static boolean getKeepCommandFile(Configuration conf) { method setKeepCommandFile (line 336) | public static void setKeepCommandFile(Configuration conf, boolean keep) { method setupPipesJob (line 340) | private static void setupPipesJob(Job job) method run (line 453) | public int run(String[] args) throws Exception { method main (line 548) | public static void main(String[] args) throws Exception { FILE: src/it/crs4/pydoop/mapreduce/pipes/TaskLog.java class TaskLog (line 71) | @InterfaceAudience.Private method getYarnAppContainerLogDir (line 84) | private static String getYarnAppContainerLogDir(){ method getMRv2LogDir (line 93) | public static String getMRv2LogDir() { method getTaskLogFile (line 97) | public static File getTaskLogFile(TaskAttemptID taskid, boolean isClea... method getRealTaskLogFileLocation (line 106) | static File getRealTaskLogFileLocation(TaskAttemptID taskid, class LogFileDetail (line 117) | private static class LogFileDetail { method getLogFileDetail (line 124) | private static LogFileDetail getLogFileDetail(TaskAttemptID taskid, method getTmpIndexFile (line 177) | private static File getTmpIndexFile(TaskAttemptID taskid, boolean isCl... method getIndexFile (line 181) | static File getIndexFile(TaskAttemptID taskid, boolean isCleanup) { method obtainLogDirOwner (line 189) | static String obtainLogDirOwner(TaskAttemptID taskid) throws IOExcepti... method getBaseLogDir (line 197) | static String getBaseLogDir() { method getAttemptDir (line 201) | static File getAttemptDir(TaskAttemptID taskid, boolean isCleanup) { method writeToIndexFile (line 209) | private static synchronized method resetPrevLengths (line 255) | private static void resetPrevLengths(String logLocation) { method syncLogs (line 262) | @SuppressWarnings("unchecked") method syncLogsShutdown (line 287) | public static synchronized void syncLogsShutdown( method syncLogs (line 303) | @SuppressWarnings("unchecked") method flushAppenders (line 322) | @SuppressWarnings("unchecked") method createLogSyncer (line 338) | public static ScheduledExecutorService createLogSyncer() { type LogName (line 369) | @InterfaceAudience.Private method LogName (line 388) | private LogName(String prefix) { method toString (line 392) | @Override class Reader (line 398) | public static class Reader extends InputStream { method Reader (line 414) | public Reader(TaskAttemptID taskid, LogName kind, method read (line 446) | @Override method read (line 456) | @Override method available (line 466) | @Override method close (line 471) | @Override method getTaskLogLength (line 485) | public static long getTaskLogLength(Configuration conf) { method captureOutAndError (line 503) | public static List captureOutAndError(List setup, method buildCommandLine (line 530) | static String buildCommandLine(List setup, List cmd, method buildDebugScriptCommandLine (line 593) | static String buildDebugScriptCommandLine(List cmd, String deb... method addCommand (line 624) | public static String addCommand(List cmd, boolean isExecutable) method getUserLogDir (line 649) | static File getUserLogDir() { method getJobDir (line 665) | public static File getJobDir(JobID jobid) { FILE: src/it/crs4/pydoop/mapreduce/pipes/TaskLogAppender.java class TaskLogAppender (line 38) | @InterfaceStability.Unstable method activateOptions (line 51) | @Override method setOptionsFromSystemProperties (line 70) | private synchronized void setOptionsFromSystemProperties() { method append (line 86) | @Override method flush (line 100) | @Override method close (line 107) | @Override method getTaskId (line 121) | public synchronized String getTaskId() { method setTaskId (line 125) | public synchronized void setTaskId(String taskId) { method getTotalLogFileSize (line 131) | public synchronized long getTotalLogFileSize() { method setTotalLogFileSize (line 135) | public synchronized void setTotalLogFileSize(long logSize) { method setIsCleanup (line 145) | public synchronized void setIsCleanup(boolean isCleanup) { method getIsCleanup (line 154) | public synchronized boolean getIsCleanup() { FILE: src/it/crs4/pydoop/mapreduce/pipes/UpwardProtocol.java type UpwardProtocol (line 29) | interface UpwardProtocol