filename
stringlengths 34
86
| repo
stringclasses 2
values | path
stringlengths 81
173
| dbytes
unknown | dbytes_len
int64 349k
1.26M
| dbytes_mb
float64 0.33
1.2
| type
stringclasses 1
value |
---|---|---|---|---|---|---|
image.class.old_test_xla_examples.TorchXLAExamplesTests.png | transformers | transformers/examples/pytorch/image.class.old_test_xla_examples.TorchXLAExamplesTests.png | null | 382,696 | 0.36 | png |
image.class.test_accelerate_examples.ExamplesTestsNoTrainer.png | transformers | transformers/examples/pytorch/image.class.test_accelerate_examples.ExamplesTestsNoTrainer.png | null | 949,697 | 0.91 | png |
image.class.test_pytorch_examples.ExamplesTests.png | transformers | transformers/examples/pytorch/image.class.test_pytorch_examples.ExamplesTests.png | null | 1,255,351 | 1.2 | png |
image.class.run_speech_recognition_seq2seq.DataCollatorSpeechSeq2SeqWithPadding.png | transformers | transformers/examples/pytorch/speech-recognition/image.class.run_speech_recognition_seq2seq.DataCollatorSpeechSeq2SeqWithPadding.png | null | 419,516 | 0.4 | png |
image.class.run_speech_recognition_ctc_adapter.DataCollatorCTCWithPadding.png | transformers | transformers/examples/pytorch/speech-recognition/image.class.run_speech_recognition_ctc_adapter.DataCollatorCTCWithPadding.png | null | 522,005 | 0.5 | png |
image.class.run_speech_recognition_ctc.DataCollatorCTCWithPadding.png | transformers | transformers/examples/pytorch/speech-recognition/image.class.run_speech_recognition_ctc.DataCollatorCTCWithPadding.png | null | 517,013 | 0.49 | png |
image.class.run_swag_no_trainer.DataCollatorForMultipleChoice.png | transformers | transformers/examples/pytorch/multiple-choice/image.class.run_swag_no_trainer.DataCollatorForMultipleChoice.png | null | 488,890 | 0.47 | png |
image.class.run_swag.DataCollatorForMultipleChoice.png | transformers | transformers/examples/pytorch/multiple-choice/image.class.run_swag.DataCollatorForMultipleChoice.png | null | 484,804 | 0.46 | png |
image.class.run_swag.DataTrainingArguments.png | transformers | transformers/examples/pytorch/multiple-choice/image.class.run_swag.DataTrainingArguments.png | null | 361,652 | 0.34 | png |
image.class.run_clip.DataTrainingArguments.png | transformers | transformers/examples/pytorch/contrastive-image-text/image.class.run_clip.DataTrainingArguments.png | null | 359,621 | 0.34 | png |
image.class.run_clip.Transform.png | transformers | transformers/examples/pytorch/contrastive-image-text/image.class.run_clip.Transform.png | null | 369,117 | 0.35 | png |
image.class.run_translation.DataTrainingArguments.png | transformers | transformers/examples/pytorch/translation/image.class.run_translation.DataTrainingArguments.png | null | 355,489 | 0.34 | png |
image.class.run_ner.DataTrainingArguments.png | transformers | transformers/examples/pytorch/token-classification/image.class.run_ner.DataTrainingArguments.png | null | 360,489 | 0.34 | png |
image.class.run_summarization.DataTrainingArguments.png | transformers | transformers/examples/pytorch/summarization/image.class.run_summarization.DataTrainingArguments.png | null | 359,016 | 0.34 | png |
image.class.run_glue.DataTrainingArguments.png | transformers | transformers/examples/pytorch/text-classification/image.class.run_glue.DataTrainingArguments.png | null | 373,934 | 0.36 | png |
image.class.run_classification.DataTrainingArguments.png | transformers | transformers/examples/pytorch/text-classification/image.class.run_classification.DataTrainingArguments.png | null | 380,342 | 0.36 | png |
image.class.run_clm.ModelArguments.png | transformers | transformers/examples/pytorch/language-modeling/image.class.run_clm.ModelArguments.png | null | 358,058 | 0.34 | png |
image.class.run_mlm.ModelArguments.png | transformers | transformers/examples/pytorch/language-modeling/image.class.run_mlm.ModelArguments.png | null | 355,804 | 0.34 | png |
image.class.run_clm.DataTrainingArguments.png | transformers | transformers/examples/pytorch/language-modeling/image.class.run_clm.DataTrainingArguments.png | null | 358,482 | 0.34 | png |
image.class.run_mlm.DataTrainingArguments.png | transformers | transformers/examples/pytorch/language-modeling/image.class.run_mlm.DataTrainingArguments.png | null | 360,218 | 0.34 | png |
image.class.run_plm.DataTrainingArguments.png | transformers | transformers/examples/pytorch/language-modeling/image.class.run_plm.DataTrainingArguments.png | null | 358,998 | 0.34 | png |
image.class.run_plm.ModelArguments.png | transformers | transformers/examples/pytorch/language-modeling/image.class.run_plm.ModelArguments.png | null | 355,196 | 0.34 | png |
image.class.run_qa.DataTrainingArguments.png | transformers | transformers/examples/pytorch/question-answering/image.class.run_qa.DataTrainingArguments.png | null | 358,984 | 0.34 | png |
image.class.run_qa_beam_search.DataTrainingArguments.png | transformers | transformers/examples/pytorch/question-answering/image.class.run_qa_beam_search.DataTrainingArguments.png | null | 367,768 | 0.35 | png |
image.class.run_seq2seq_qa.DataTrainingArguments.png | transformers | transformers/examples/pytorch/question-answering/image.class.run_seq2seq_qa.DataTrainingArguments.png | null | 366,543 | 0.35 | png |
image.class.trainer_qa.QuestionAnsweringTrainer.png | transformers | transformers/examples/pytorch/question-answering/image.class.trainer_qa.QuestionAnsweringTrainer.png | null | 441,618 | 0.42 | png |
image.class.trainer_seq2seq_qa.QuestionAnsweringSeq2SeqTrainer.png | transformers | transformers/examples/pytorch/question-answering/image.class.trainer_seq2seq_qa.QuestionAnsweringSeq2SeqTrainer.png | null | 463,275 | 0.44 | png |
image.class.run_audio_classification.ModelArguments.png | transformers | transformers/examples/pytorch/audio-classification/image.class.run_audio_classification.ModelArguments.png | null | 366,376 | 0.35 | png |
image.class.run_wav2vec2_pretraining_no_trainer.DataCollatorForWav2Vec2Pretraining.png | transformers | transformers/examples/pytorch/speech-pretraining/image.class.run_wav2vec2_pretraining_no_trainer.DataCollatorForWav2Vec2Pretraining.png | null | 633,270 | 0.6 | png |
image.class.run_mae.DataTrainingArguments.png | transformers | transformers/examples/pytorch/image-pretraining/image.class.run_mae.DataTrainingArguments.png | null | 375,150 | 0.36 | png |
image.class.run_mim.DataTrainingArguments.png | transformers | transformers/examples/pytorch/image-pretraining/image.class.run_mim.DataTrainingArguments.png | null | 371,731 | 0.35 | png |
image.class.run_mim.MaskGenerator.png | transformers | transformers/examples/pytorch/image-pretraining/image.class.run_mim.MaskGenerator.png | null | 387,535 | 0.37 | png |
image.class.run_mim_no_trainer.MaskGenerator.png | transformers | transformers/examples/pytorch/image-pretraining/image.class.run_mim_no_trainer.MaskGenerator.png | null | 394,707 | 0.38 | png |
image.class.run_image_classification.DataTrainingArguments.png | transformers | transformers/examples/pytorch/image-classification/image.class.run_image_classification.DataTrainingArguments.png | null | 384,012 | 0.37 | png |
image.class.run_generation._ModelFallbackWrapper.png | transformers | transformers/examples/pytorch/text-generation/image.class.run_generation._ModelFallbackWrapper.png | null | 538,736 | 0.51 | png |
image.class.run_semantic_segmentation.ReduceLabels.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.ReduceLabels.png | null | 352,824 | 0.34 | png |
image.class.run_semantic_segmentation.DataTrainingArguments.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.DataTrainingArguments.png | null | 377,257 | 0.36 | png |
image.class.run_semantic_segmentation.PILToTensor.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.PILToTensor.png | null | 349,301 | 0.33 | png |
image.class.run_semantic_segmentation.Compose.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.Compose.png | null | 370,332 | 0.35 | png |
image.class.run_semantic_segmentation.RandomCrop.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.RandomCrop.png | null | 373,517 | 0.36 | png |
image.class.run_semantic_segmentation.RandomResize.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.RandomResize.png | null | 373,889 | 0.36 | png |
image.class.run_semantic_segmentation.Resize.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.Resize.png | null | 363,917 | 0.35 | png |
image.class.run_semantic_segmentation.Identity.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.Identity.png | null | 362,738 | 0.35 | png |
image.class.run_semantic_segmentation.RandomHorizontalFlip.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.RandomHorizontalFlip.png | null | 380,424 | 0.36 | png |
image.class.run_semantic_segmentation.Normalize.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.Normalize.png | null | 365,257 | 0.35 | png |
image.class.run_semantic_segmentation.ConvertImageDtype.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.ConvertImageDtype.png | null | 381,396 | 0.36 | png |
image.class.run_semantic_segmentation_no_trainer.PILToTensor.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.PILToTensor.png | null | 355,473 | 0.34 | png |
image.class.run_semantic_segmentation_no_trainer.ReduceLabels.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.ReduceLabels.png | null | 356,414 | 0.34 | png |
image.class.run_semantic_segmentation_no_trainer.Normalize.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.Normalize.png | null | 371,564 | 0.35 | png |
image.class.run_semantic_segmentation_no_trainer.ConvertImageDtype.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.ConvertImageDtype.png | null | 386,738 | 0.37 | png |
image.class.run_semantic_segmentation_no_trainer.RandomHorizontalFlip.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.RandomHorizontalFlip.png | null | 385,779 | 0.37 | png |
image.class.run_semantic_segmentation_no_trainer.RandomCrop.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.RandomCrop.png | null | 377,654 | 0.36 | png |
image.class.run_semantic_segmentation_no_trainer.RandomResize.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.RandomResize.png | null | 379,863 | 0.36 | png |
image.class.run_semantic_segmentation_no_trainer.Identity.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.Identity.png | null | 371,312 | 0.35 | png |
image.class.run_semantic_segmentation_no_trainer.Compose.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.Compose.png | null | 376,378 | 0.36 | png |
image.class.run_semantic_segmentation_no_trainer.Resize.png | transformers | transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.Resize.png | null | 372,677 | 0.36 | png |
image.class.test_xla_examples.TorchXLAExamplesTests.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/image.class.test_xla_examples.TorchXLAExamplesTests.png | null | 401,421 | 0.38 | png |
image.class.test_accelerate_examples.ExamplesTestsNoTrainer.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/image.class.test_accelerate_examples.ExamplesTestsNoTrainer.png | null | 960,106 | 0.92 | png |
image.class.test_pytorch_examples.ExamplesTests.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/image.class.test_pytorch_examples.ExamplesTests.png | null | 1,195,227 | 1.14 | png |
image.class.run_wav2vec2_pretraining_no_trainer.DataCollatorForWav2Vec2Pretraining.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/speech-pretraining/image.class.run_wav2vec2_pretraining_no_trainer.DataCollatorForWav2Vec2Pretraining.png | null | 567,658 | 0.54 | png |
image.class.run_swag_no_trainer.DataCollatorForMultipleChoice.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/multiple-choice/image.class.run_swag_no_trainer.DataCollatorForMultipleChoice.png | null | 509,651 | 0.49 | png |
image.class.run_swag.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/multiple-choice/image.class.run_swag.DataTrainingArguments.png | null | 377,922 | 0.36 | png |
image.class.run_swag.DataCollatorForMultipleChoice.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/multiple-choice/image.class.run_swag.DataCollatorForMultipleChoice.png | null | 501,624 | 0.48 | png |
image.class.run_speech_recognition_ctc.DataCollatorCTCWithPadding.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/speech-recognition/image.class.run_speech_recognition_ctc.DataCollatorCTCWithPadding.png | null | 533,895 | 0.51 | png |
image.class.run_speech_recognition_seq2seq.DataCollatorSpeechSeq2SeqWithPadding.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/speech-recognition/image.class.run_speech_recognition_seq2seq.DataCollatorSpeechSeq2SeqWithPadding.png | null | 428,427 | 0.41 | png |
image.class.run_clip.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/contrastive-image-text/image.class.run_clip.DataTrainingArguments.png | null | 380,355 | 0.36 | png |
image.class.run_clip.Transform.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/contrastive-image-text/image.class.run_clip.Transform.png | null | 389,604 | 0.37 | png |
image.class.run_glue.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/text-classification/image.class.run_glue.DataTrainingArguments.png | null | 394,805 | 0.38 | png |
image.class.run_plm.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/language-modeling/image.class.run_plm.DataTrainingArguments.png | null | 378,025 | 0.36 | png |
image.class.run_plm.ModelArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/language-modeling/image.class.run_plm.ModelArguments.png | null | 376,954 | 0.36 | png |
image.class.run_mlm.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/language-modeling/image.class.run_mlm.DataTrainingArguments.png | null | 377,886 | 0.36 | png |
image.class.run_mlm.ModelArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/language-modeling/image.class.run_mlm.ModelArguments.png | null | 376,835 | 0.36 | png |
image.class.run_clm.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/language-modeling/image.class.run_clm.DataTrainingArguments.png | null | 380,400 | 0.36 | png |
image.class.run_clm.ModelArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/language-modeling/image.class.run_clm.ModelArguments.png | null | 374,881 | 0.36 | png |
image.class.run_qa.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/question-answering/image.class.run_qa.DataTrainingArguments.png | null | 380,438 | 0.36 | png |
image.class.run_qa_beam_search.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/question-answering/image.class.run_qa_beam_search.DataTrainingArguments.png | null | 385,321 | 0.37 | png |
image.class.run_seq2seq_qa.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/question-answering/image.class.run_seq2seq_qa.DataTrainingArguments.png | null | 386,176 | 0.37 | png |
image.class.trainer_seq2seq_qa.QuestionAnsweringSeq2SeqTrainer.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/question-answering/image.class.trainer_seq2seq_qa.QuestionAnsweringSeq2SeqTrainer.png | null | 469,127 | 0.45 | png |
image.class.trainer_qa.QuestionAnsweringTrainer.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/question-answering/image.class.trainer_qa.QuestionAnsweringTrainer.png | null | 448,273 | 0.43 | png |
image.class.run_mae.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/image-pretraining/image.class.run_mae.DataTrainingArguments.png | null | 389,426 | 0.37 | png |
image.class.run_mim.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/image-pretraining/image.class.run_mim.DataTrainingArguments.png | null | 390,854 | 0.37 | png |
image.class.run_mim.MaskGenerator.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/image-pretraining/image.class.run_mim.MaskGenerator.png | null | 405,121 | 0.39 | png |
image.class.run_audio_classification.ModelArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/audio-classification/image.class.run_audio_classification.ModelArguments.png | null | 381,679 | 0.36 | png |
image.class.run_summarization.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/summarization/image.class.run_summarization.DataTrainingArguments.png | null | 381,661 | 0.36 | png |
image.class.run_translation.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/translation/image.class.run_translation.DataTrainingArguments.png | null | 376,080 | 0.36 | png |
image.class.run_ner.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/token-classification/image.class.run_ner.DataTrainingArguments.png | null | 380,076 | 0.36 | png |
image.class.run_image_classification.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/image-classification/image.class.run_image_classification.DataTrainingArguments.png | null | 400,086 | 0.38 | png |
image.class.run_semantic_segmentation_no_trainer.PILToTensor.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.PILToTensor.png | null | 376,100 | 0.36 | png |
image.class.run_semantic_segmentation_no_trainer.ReduceLabels.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.ReduceLabels.png | null | 377,195 | 0.36 | png |
image.class.run_semantic_segmentation_no_trainer.ConvertImageDtype.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.ConvertImageDtype.png | null | 407,341 | 0.39 | png |
image.class.run_semantic_segmentation_no_trainer.Normalize.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.Normalize.png | null | 388,377 | 0.37 | png |
image.class.run_semantic_segmentation_no_trainer.Compose.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.Compose.png | null | 392,349 | 0.37 | png |
image.class.run_semantic_segmentation_no_trainer.Identity.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.Identity.png | null | 385,824 | 0.37 | png |
image.class.run_semantic_segmentation_no_trainer.Resize.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.Resize.png | null | 388,817 | 0.37 | png |
image.class.run_semantic_segmentation_no_trainer.RandomResize.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.RandomResize.png | null | 398,581 | 0.38 | png |
image.class.run_semantic_segmentation_no_trainer.RandomCrop.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.RandomCrop.png | null | 394,201 | 0.38 | png |
image.class.run_semantic_segmentation_no_trainer.RandomHorizontalFlip.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation_no_trainer.RandomHorizontalFlip.png | null | 404,820 | 0.39 | png |
image.class.run_semantic_segmentation.ReduceLabels.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.ReduceLabels.png | null | 370,869 | 0.35 | png |
image.class.run_semantic_segmentation.PILToTensor.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.PILToTensor.png | null | 369,866 | 0.35 | png |
image.class.run_semantic_segmentation.DataTrainingArguments.png | H2O | H2O/h2o_flexgen/benchmark/third_party/transformers/examples/pytorch/semantic-segmentation/image.class.run_semantic_segmentation.DataTrainingArguments.png | null | 395,809 | 0.38 | png |