This repository contains code released by Google Research.

Overview

Google Research

This repository contains code released by Google Research.

All datasets in this repository are released under the CC BY 4.0 International license, which can be found here: https://creativecommons.org/licenses/by/4.0/legalcode. All source files in this repository are released under the Apache 2.0 license, the text of which can be found in the LICENSE file.


Because the repo is large, we recommend you download only the subdirectory of interest:

SUBDIR=foo
svn export https://github.com/google-research/google-research/trunk/$SUBDIR

If you'd like to submit a pull request, you'll need to clone the repository; we recommend making a shallow clone (without history).

git clone [email protected]:google-research/google-research.git --depth=1

Disclaimer: This is not an official Google product.

Comments
  • depth_from_video_in_the_wild: not able to reproduce the result

    depth_from_video_in_the_wild: not able to reproduce the result

    @gariel-google Dear author, Thanks for sharing the source code of the paper.
    I was trying to reproduce the result of the paper using your code. However, with your default setting (batch size=4, learning_rate=0.0002, etc.) training from scratch, the result I got it's quite far from what you stated in the paper (Abs Rel 0.147 for the best checkpoint within around 370k-th step vs 0.128 in the paper). For your information, I am using the evaluation code from sfmlearner as what struct2depth does.
    Therefore, may I know what's setting for obtaining the paper's result? Or is there anything critical part missing in the current released code (maybe pretrained checkpoint for example)?
    Thank you in advance.

    opened by liyingliu 63
  • about coltran, I have some questions.

    about coltran, I have some questions.

    I have performed a grayscale image coloring test based on the pre-trained model you provided @MechCoder**[(https://storage.cloud.google.com/gresearch/coltran/coltran.zip.)**] Why are the results all noisy images? I have executed the custom_colorize.py in sequence according to the file description you provided. (2) (3) (4) (5) (6) (7) (8) (9) (10) (1)

    opened by hufeng0912 36
  • Colorization Transformer: Single GrayScale Image Inference Script?

    Colorization Transformer: Single GrayScale Image Inference Script?

    @MechCoder It seems that Colorization Transformer didn't have a script for single grayscale image inference.

    I am not familiar with the TF2.0, thus there are also some questions for this open source code:

    1. When i try to adapt the sample.py for single image inference, i found it will download the imagenet2012 dataset. Is this nessasary for sample.py?
    2. Is there any difference between ”Sampling“ and ”inference“ ?
    opened by ccpocker 35
  • depth_from_video_in_the_wild: image size for pretrained models

    depth_from_video_in_the_wild: image size for pretrained models

    Hi @gariel-google, are the models that you provide trained on images 416x128? When I tried inference with other resolutions it doesn't work well at all.

    If it's indeed 416x128, have you tried training with higher resolutions? I know some previous work use 416x128 for training, but recently most methods use higher resolutions and experiments have demonstrated higher resolutions lead to better results. Is it something related to the GPU memory issue?

    opened by kwea123 31
  • About the coltran ,I want ask a idiot question?

    About the coltran ,I want ask a idiot question?

    Absolutely ,you did a excellent job! I'm not familiar tf and in your paper figure 2 show the input image only have grayscale image,so I want to ask that when training ,Is there only a gray image input? Hope you can reply me about this idiot question,Thanks

    opened by meiguoofa 24
  • tf_trees gives segmentation fault

    tf_trees gives segmentation fault

    I have been trying to use tf_trees and although the compilation goes without any warning in a amazon linux (built on top of Fedora if I understood correctly) the demo.py fails with segmentation fault. Here is the specification of the machine

    NAME="Amazon Linux AMI"
    VERSION="2018.03"
    ID="amzn"
    ID_LIKE="rhel fedora"
    VERSION_ID="2018.03"
    PRETTY_NAME="Amazon Linux AMI 2018.03"
    ANSI_COLOR="0;33"
    CPE_NAME="cpe:/o:amazon:linux:2018.03:ga"
    HOME_URL="http://aws.amazon.com/amazon-linux-ami/"
    Amazon Linux AMI release 2018.03
    

    I cannot compile the same on a colab (which is an ubuntu environment).

    DISTRIB_ID=Ubuntu
    DISTRIB_RELEASE=18.04
    DISTRIB_CODENAME=bionic
    DISTRIB_DESCRIPTION="Ubuntu 18.04.5 LTS"
    NAME="Ubuntu"
    VERSION="18.04.5 LTS (Bionic Beaver)"
    ID=ubuntu
    ID_LIKE=debian
    PRETTY_NAME="Ubuntu 18.04.5 LTS"
    VERSION_ID="18.04"
    HOME_URL="https://www.ubuntu.com/"
    SUPPORT_URL="https://help.ubuntu.com/"
    BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
    PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
    VERSION_CODENAME=bionic
    UBUNTU_CODENAME=bionic
    

    when I try to compile it prints the following and does not produce the .so file

    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/partial_tensor_shape.h:20:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/attr_value_util.h:23,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/node_def_util.h:23,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/full_type_util.h:24,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op.h:24,
                     from neural_trees_ops.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_shape.h:305:22: warning: ‘tensorflow::int64’ is deprecated: Use int64_t instead. [-Wdeprecated-declarations]
       gtl::InlinedVector<int64, 4> dim_sizes() const;
                          ^~~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/types.h:31:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/logging.h:20,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/errors.h:24,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/lib/core/errors.h:19,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_shape.h:23,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/partial_tensor_shape.h:20,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/attr_value_util.h:23,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/node_def_util.h:23,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/full_type_util.h:24,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op.h:24,
                     from neural_trees_ops.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/integral_types.h:29:63: note: declared here
     [[deprecated("Use int64_t instead.")]] typedef ::std::int64_t int64;
                                                                   ^~~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor.h:25:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/attr_value_util.h:24,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/node_def_util.h:23,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/full_type_util.h:24,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op.h:24,
                     from neural_trees_ops.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h: In member function ‘void tensorflow::internal::MaybeWith32BitIndexingImpl<Eigen::GpuDevice>::operator()(Func, Args&& ...) const’:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h:176:25: error: use of ‘auto’ in lambda parameter declaration only available with -std=c++14 or -std=gnu++14
         auto all = [](const auto&... bool_vals) {
                             ^~~~
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h:176:34: error: expansion pattern ‘const int&’ contains no argument packs
         auto all = [](const auto&... bool_vals) {
                                      ^~~~~~~~~
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h: In lambda function:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h:177:22: error: ‘bool_vals’ was not declared in this scope
           for (bool b : {bool_vals...}) {
                          ^~~~~~~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/notification.h:27:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/lib/core/notification.h:21,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/cancellation.h:22,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:27,
                     from neural_trees_helpers.h:17,
                     from neural_trees_kernels.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/notification.h:61:65: warning: ‘int64’ is deprecated [-Wdeprecated-declarations]
                                                  int64 timeout_in_us);
                                                                     ^
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/notification.h:62:58: warning: ‘int64’ is deprecated [-Wdeprecated-declarations]
       bool WaitForNotificationWithTimeout(int64 timeout_in_us) {
                                                              ^
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/notification.h:81:63: warning: ‘int64’ is deprecated [-Wdeprecated-declarations]
                                                int64 timeout_in_us) {
                                                                   ^
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor.h:24:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/device_base.h:26,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:29,
                     from neural_trees_helpers.h:17,
                     from neural_trees_kernels.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_shape.h:305:22: warning: ‘tensorflow::int64’ is deprecated: Use int64_t instead. [-Wdeprecated-declarations]
       gtl::InlinedVector<int64, 4> dim_sizes() const;
                          ^~~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/types.h:31:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/numeric_types.h:27,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/allocator.h:26,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:26,
                     from neural_trees_helpers.h:17,
                     from neural_trees_kernels.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/integral_types.h:29:63: note: declared here
     [[deprecated("Use int64_t instead.")]] typedef ::std::int64_t int64;
                                                                   ^~~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor.h:25:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/device_base.h:26,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:29,
                     from neural_trees_helpers.h:17,
                     from neural_trees_kernels.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h: In member function ‘void tensorflow::internal::MaybeWith32BitIndexingImpl<Eigen::GpuDevice>::operator()(Func, Args&& ...) const’:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h:176:25: error: use of ‘auto’ in lambda parameter declaration only available with -std=c++14 or -std=gnu++14
         auto all = [](const auto&... bool_vals) {
                             ^~~~
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h:176:34: error: expansion pattern ‘const int&’ contains no argument packs
         auto all = [](const auto&... bool_vals) {
                                      ^~~~~~~~~
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h: In lambda function:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h:177:22: error: ‘bool_vals’ was not declared in this scope
           for (bool b : {bool_vals...}) {
                          ^~~~~~~~~
    neural_trees_kernels.cc: In member function ‘virtual void tensorflow::NTComputeInputAndInternalParamsGradientsOp::Compute(tensorflow::OpKernelContext*)’:
    neural_trees_kernels.cc:141:19: warning: ‘tensorflow::int64’ is deprecated: Use int64_t instead. [-Wdeprecated-declarations]
           const int64 cost = 10000 * std::log10(input_dim) * std::log2(num_leaves);
                       ^~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/types.h:31:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/numeric_types.h:27,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/allocator.h:26,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:26,
                     from neural_trees_helpers.h:17,
                     from neural_trees_kernels.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/integral_types.h:29:63: note: declared here
     [[deprecated("Use int64_t instead.")]] typedef ::std::int64_t int64;
                                                                   ^~~~~
    neural_trees_kernels.cc: In member function ‘virtual void tensorflow::NTComputeOutputOp::Compute(tensorflow::OpKernelContext*)’:
    neural_trees_kernels.cc:274:19: warning: ‘tensorflow::int64’ is deprecated: Use int64_t instead. [-Wdeprecated-declarations]
           const int64 cost = 10000 * std::log10(input_dim) * std::log2(num_leaves);
                       ^~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/types.h:31:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/numeric_types.h:27,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/allocator.h:26,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:26,
                     from neural_trees_helpers.h:17,
                     from neural_trees_kernels.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/integral_types.h:29:63: note: declared here
     [[deprecated("Use int64_t instead.")]] typedef ::std::int64_t int64;
                                                                   ^~~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/notification.h:27:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/lib/core/notification.h:21,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/cancellation.h:22,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:27,
                     from neural_trees_helpers.h:17,
                     from neural_trees_helpers.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/notification.h:61:65: warning: ‘int64’ is deprecated [-Wdeprecated-declarations]
                                                  int64 timeout_in_us);
                                                                     ^
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/notification.h:62:58: warning: ‘int64’ is deprecated [-Wdeprecated-declarations]
       bool WaitForNotificationWithTimeout(int64 timeout_in_us) {
                                                              ^
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/notification.h:81:63: warning: ‘int64’ is deprecated [-Wdeprecated-declarations]
                                                int64 timeout_in_us) {
                                                                   ^
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor.h:24:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/device_base.h:26,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:29,
                     from neural_trees_helpers.h:17,
                     from neural_trees_helpers.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_shape.h:305:22: warning: ‘tensorflow::int64’ is deprecated: Use int64_t instead. [-Wdeprecated-declarations]
       gtl::InlinedVector<int64, 4> dim_sizes() const;
                          ^~~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/types.h:31:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/numeric_types.h:27,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/allocator.h:26,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:26,
                     from neural_trees_helpers.h:17,
                     from neural_trees_helpers.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/platform/default/integral_types.h:29:63: note: declared here
     [[deprecated("Use int64_t instead.")]] typedef ::std::int64_t int64;
                                                                   ^~~~~
    In file included from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor.h:25:0,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/device_base.h:26,
                     from /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/op_kernel.h:29,
                     from neural_trees_helpers.h:17,
                     from neural_trees_helpers.cc:15:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h: In member function ‘void tensorflow::internal::MaybeWith32BitIndexingImpl<Eigen::GpuDevice>::operator()(Func, Args&& ...) const’:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h:176:25: error: use of ‘auto’ in lambda parameter declaration only available with -std=c++14 or -std=gnu++14
         auto all = [](const auto&... bool_vals) {
                             ^~~~
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h:176:34: error: expansion pattern ‘const int&’ contains no argument packs
         auto all = [](const auto&... bool_vals) {
                                      ^~~~~~~~~
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h: In lambda function:
    /usr/local/lib/python3.7/dist-packages/tensorflow/include/tensorflow/core/framework/tensor_types.h:177:22: error: ‘bool_vals’ was not declared in this scope
           for (bool b : {bool_vals...}) {
                          ^~~~~~~~~
    

    What can I do? I would have loved to try it, modify it, and use it in my project.

    Please help

    opened by rcshubhadeep 18
  • How the postprocessing in object detection works?

    How the postprocessing in object detection works?

    Hello all,

    Thank you for your wonderful work on Tensorflow 3D for lidar point cloud.

    I have a question regarding post processing in object detection post processing.

    what is the input shape to the postprocessing? I mean, for an example, usually we get [x,y,z,h,w,l,r] in kitti data set as a output of the model. is input to the postprocessing same as the kitti based object detection model?

    For an example Could anyone help me to understand?

    opened by mariya12290 18
  • Build did NOT complete successfully - tf3d

    Build did NOT complete successfully - tf3d

    I'm following line by line this guide for working with Sparse convolution but when I need to do the following command

    bazel run sparse_conv_ops_py_test --experimental_repo_remote_exec

    I get the following error:

    ERROR: Skipping 'sparse_conv_ops_py_test': couldn't determine target from filename 'sparse_conv_ops_py_test'
    WARNING: Target pattern parsing failed.
    ERROR: couldn't determine target from filename 'sparse_conv_ops_py_test'
    INFO: Elapsed time: 0.222s
    INFO: 0 processes.
    FAILED: Build did NOT complete successfully (0 packages loaded)
    FAILED: Build did NOT complete successfully (0 packages loaded)
    

    I'm running this inside my docker image on my Mac OS. I'm running docker run without runtime nvidia as my mac does not have it

    docker run --privileged -it -v ${MYFOLDER}:/working_dir -w /working_dir tensorflow/tensorflow:2.3.0-custom-op-gpu-ubuntu16

    When I run configure.sh I get the following which seems to be ok

    root@562522a0cd6e:/working_dir/tf3d/ops# ./configure.sh
    Does the pip package have tag manylinux2010 (usually the case for nightly release after Aug 1, 2019, or official releases past 1.14.0)?. Y or enter for manylinux2010, N for manylinux1. [Y/n] y
    Are you building against TensorFlow 2.1(including RCs) or newer?[Y/n] y
    Build against TensorFlow 2.1 or newer.
    Using installed tensorflow
    2021-03-17 10:32:03.125008: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1
    2021-03-17 10:32:09.258489: I tensorflow/stream_executor/platform/default/dso_loader.cc:48] Successfully opened dynamic library libcudart.so.10.1
    

    Also i had to create a WORKSPACE file before running the bazel command in the same directory

    opened by FrancescoMandru 16
  • OOM when jax.jit is used to compile and execute mu2Net train_step function

    OOM when jax.jit is used to compile and execute mu2Net train_step function

    @agesmundo, hope to receive some reply, thanks.

    • Duplicate issues: #788 and #4528 are not suitable for this case.

    • How to reproduce the bug: [1] Just run mu2Net on 8gpus A100, use BENCHMARK = 'ViT large / Chars benchmark' [2] OOM error will occur when train_step function compiled by jax.jit is executed. [3] The A100 have sufficent 80GiB memory per gpu, i use 8gpus. My cpu has 256g memory and 112+ cores. [4] I can't understand why the executable needs to preallocate 114.44GiB temp allocation, though the seed ViT model is just 300M. [5]It's useless to set any env variable about jax memory allocation

    • Model hyperparameters:

    def get_exp_config_large(benchmark_string_id):
      exp_config = ConfigDict()
      exp_config.experiment_name = EXPERIMENT_NAME
      exp_config.experiments_root_dir = EXPERIMENTS_ROOT_DIR
      # Cap to 1/10th of imagenet train set size to have similar ratio of exps reported in:
      # https://arxiv.org/abs/2106.10270
      exp_config.num_train_examples_between_validations_max = 128_116
      exp_config.num_validations_per_path_training = 4
      exp_config.num_validation_examples_max = 10_000
      # Fit HBM memory: TPUv4 megacore=64, TPUv3=32.
      exp_config.batch_size = 64
      exp_config.num_task_iters = 1
      # Assuming TPUv4 32 cores * 4 generations.
      exp_config.num_samples_per_task = 32 * 4
      exp_config.mutate_adapters = False
      exp_config.force_finetune_components = ['encoder_norm']
      # Population policy params:
      exp_config.policy_class = 'PPDecay'
      exp_config.policy_kwargs = {}
      # Scorer params:
      exp_config.scorer_class = 'ScorerDecay'
      exp_config.scorer_kwargs = dict(
          base=1.0,
          num_params=303_303_682,  # Params in L/16
          )
      # Seed models params:
      exp_config.load_rand_init = False
      exp_config.load_vit_checkpoint = True
      exp_config.load_vit_checkpoint_query = 'name=="L/16" and ds=="i21k" and aug=="medium2" and wd==0.03 and sd==0.1'
      exp_config.load_experiment = False
      exp_config.load_experiment_dir = ''
      set_continue_configs(exp_config)
    
      # Hyperparameters:
      max_num_layers = get_max_num_layers(exp_config.load_vit_checkpoint_query)
      exp_config.models_default_hparams = {
          '_mu_': 0.2,
          'num_classes': 1,
          'adapter_layers': '',
          'num_layers': max_num_layers,
          'adapter_dim': 16,
          'opt_lr': 0.01,
          'opt_lr_schedule': 'cosine',
          'opt_lr_warmup_ratio': 0.05,
          'opt_momentum': 0.9,
          'opt_nesterov': False,
          'ds_image_size': 384,
          'ds_crop': True,
          'ds_area_range_min': 0.05,
          'ds_aspect_ratio_range_min': 0.75,
          'ds_flip_left_right': True,
          'ds_brightness_delta': 0.0,
          'ds_contrast_delta': 0.0,
          'ds_saturation_delta': 0.0,
          'ds_hue_delta': 0.0,
      }
    
    • Core code:
    @partial(jax.jit, static_argnames=['model', 'optimizer'], donate_argnums=[0, 2])
    def train_step(params, fixed_params, opt_state, images, labels, model, optimizer):
      def loss_fn(params, fixed_params, images, labels):
        logits = model.apply({'params': format_params(params, fixed_params)},
                             images, train=USE_DROPOUT)
        labels = jax.nn.one_hot(labels, logits.shape[-1])
        return -jnp.mean(jnp.sum(labels * nn.log_softmax(logits), axis=-1))
      grads = jax.grad(loss_fn)(params, fixed_params, images, labels)
      updates, opt_state = optimizer.update(grads, opt_state, params=params)
      params = optax.apply_updates(params, updates)
      return params, opt_state
    
    def train_loop(paths, ds_train, ds_validation, devices, exp_config):
      global LOOP_START
      timing = {'start_time': time.time(),
                'start_time_loop': LOOP_START}
      task = paths[0].task
      # The following values should be shared by all paths in this generation batch.
      for path in paths:
        assert task == path.task
        assert paths[0].hparams['ds_image_size'] == path.hparams['ds_image_size']
    
      gc.collect()
    
      # Compile.
      compile_train_batches_arr = jax.device_put_replicated(
          get_sample_batch(
            paths[0].hparams['ds_image_size'],
            task.train_batch_size),
          devices)
      compile_eval_batches_arr = jax.device_put_replicated(
          get_sample_batch(
              paths[0].hparams['ds_image_size'],
              task.validation_batch_size),
          devices)
    
      for p_id, path in enumerate(paths):
        if VERBOSE:
          print('Parent')
          print(prp(path.parent))
          print(prp(path))
        path.device_id = p_id % len(devices)
        path.device = devices[path.device_id]
        print("path:", p_id, "device:", path.device)
        path.optimizer = path.get_optimizer()
        path.optimizer_init_fn = jax.jit(path.optimizer.init, device=path.device)
        path.best_params_local = None
        path.best_opt_state_local = None
        path.best_quality = None
        path.best_score = path.parent.score() if path.task is path.parent.task else -np.inf
        path.evals = []
    
        # Launch parallel compilation of eval and train step functions.
        params_local = path.get_trainable_params()
        check_is_local(params_local)
        path.compile_params_device = jax.device_put(params_local, path.device)
        path.compile_fixed_params_device = jax.device_put(
            path.get_fixed_params(),
            path.device)
        path.compile_train = Thread(
            target=train_step,
            args=(path.compile_params_device,
                  path.compile_fixed_params_device,
                  path.optimizer_init_fn(params_local),
                  compile_train_batches_arr['image'][path.device_id],
                  compile_train_batches_arr['label'][path.device_id],
                  path.model,
                  path.optimizer))
        path.compile_eval = Thread(
            target=eval_step,
            args=(format_params(
                      path.compile_params_device,
                      path.compile_fixed_params_device),
                  compile_eval_batches_arr['image'][path.device_id],
                  compile_eval_batches_arr['label'][path.device_id],
                  path.model))
        path.compile_eval.start()
    
      for path in paths:
        path.compile_eval.join()
        del path.compile_eval
        timing['end_compile_eval'] = time.time()
        path.compile_train.start()
      del compile_eval_batches_arr
    
      for path in paths:
        path.compile_train.join()
        del path.compile_train
        del path.compile_params_device
        del path.compile_fixed_params_device
        timing['end_compile'] = time.time()
      del compile_train_batches_arr
    
      gc.collect()
    
      # Parameter transfer.
      for path in paths:
        path.params_device = jax.device_put(
            path.get_trainable_params(),
            path.device)
        path.fixed_params_device = jax.device_put(
            path.get_fixed_params(),
            path.device)
        path.opt_state_device = path.optimizer_init_fn(path.params_device)
        # Set opt state.
        for c in path.components:
          if c.is_trainable():
            assert c.name in path.opt_state_device[1][0].trace.keys()
            if c.opt_state is not None:
              path.opt_state_device = (
                  path.opt_state_device[0],
                  (optax.TraceState(
                      trace=path.opt_state_device[1][0].trace.copy(
                          {c.name: jax.device_put(c.opt_state,
                                                  path.device)})),
                   path.opt_state_device[1][1]
                   )
              )
        check_is_on_device(path.opt_state_device, path.device)
    
      iter_ds_validation = iter(ds_validation)
      # TRAIN
      for t_step, train_batch in zip(
          range(exp_config.num_validations_per_path_training
                * task.num_train_batches_between_validations),
          ds_train,
      ):
        train_batch_arr = jax.device_put_replicated(train_batch, devices)
        for p_id, path in enumerate(paths):
          if t_step == 0:
            timing['end_prep'] = time.time()
            t_step_0_time = time.time()
          train_step_start = time.time()
          path.params_device, path.opt_state_device = train_step(
              path.params_device,
              path.fixed_params_device,
              path.opt_state_device,
              train_batch_arr['image'][path.device_id],
              train_batch_arr['label'][path.device_id],
              path.model,
              path.optimizer)
          if t_step == 0 and time.time() - t_step_0_time > 1:
            print(f'WARNING: First train step took: {time.time()-t_step_0_time:.2f} s')
        del train_batch, train_batch_arr
    
        # EVAL
        # ...
    
    • Full error messages/tracebacks:
    Exception in thread Thread-14:
    Traceback (most recent call last):
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/threading.py", line 890, in _bootstrap
        self._bootstrap_inner()
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/threading.py", line 926, in _bootstrap_inner
        self.run()
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/threading.py", line 870, in run
        self._target(*self._args, **self._kwargs)
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/site-packages/jax/_src/traceback_util.py", line 162, in reraise_with_filtered_traceback
        return fun(*args, **kwargs)
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/site-packages/jax/_src/api.py", line 476, in cache_miss
        donated_invars=donated_invars, inline=inline, keep_unused=keep_unused)
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/site-packages/jax/core.py", line 1765, in bind
        return call_bind(self, fun, *args, **params)
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/site-packages/jax/core.py", line 1781, in call_bind
        outs = top_trace.process_call(primitive, fun_, tracers, params)
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/site-packages/jax/core.py", line 678, in process_call
        return primitive.impl(f, *tracers, **params)
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/site-packages/jax/_src/dispatch.py", line 185, in _xla_call_impl
        return compiled_fun(*args)
      File "/mnt/lustre/liujun1/.conda/envs/muNet/lib/python3.7/site-packages/jax/_src/dispatch.py", line 615, in _execute_compiled
        out_bufs_flat = compiled.execute(input_bufs_flat)
    jax._src.traceback_util.UnfilteredStackTrace: jaxlib.xla_extension.XlaRuntimeError: RESOURCE_EXHAUSTED: Out of memory while trying to allocate 122875791936 bytes.
    BufferAssignment OOM Debugging.
    BufferAssignment stats:
                 parameter allocation:    1.43GiB
                  constant allocation:         8B
            maybe_live_out allocation:  581.19MiB
         preallocated temp allocation:  114.44GiB
      preallocated temp fragmentation:  146.50MiB (0.13%)
                     total allocation:  115.86GiB
                  total fragmentation:  146.53MiB (0.12%)
    Peak buffers:
            Buffer 1:
                    Size: 1.27GiB
                    XLA Label: custom-call
                    Shape: f32[64,16,577,577]
                    ==========================
    
            Buffer 2:
                    Size: 1.27GiB
                    XLA Label: custom-call
                    Shape: f32[64,16,577,577]
                    ==========================
    
            Buffer 3:
                    Size: 1.27GiB
                    XLA Label: custom-call
                    Shape: f32[64,16,577,577]
                    ==========================
    
            Buffer 4:
                    Size: 1.27GiB
                    XLA Label: custom-call
                    Shape: f32[64,16,577,577]
                    ==========================
    
            Buffer 5:
                    Size: 1.27GiB
                    XLA Label: custom-call
                    Shape: f32[64,16,577,577]
                    ==========================
    
            Buffer 6:
                    Size: 1.27GiB
                    XLA Label: custom-call
                    Shape: f32[64,16,577,577]
                    ==========================
    
    
    • environment: python: 3.7.13 jax: 0.3.14/0.3.13 jaxlib: 0.3.14+cuda11.cudnn82/0.3.10+cuda11.cudnn82

    • python packages: absl-py 1.1.0 aqtp 0.0.7 astunparse 1.6.3 cachetools 5.2.0 certifi 2022.6.15 charset-normalizer 2.0.12 chex 0.1.3 cloudpickle 2.1.0 clu 0.0.3 colorama 0.4.5 commonmark 0.9.1 contextlib2 21.6.0 cycler 0.11.0 dacite 1.6.0 decorator 5.1.1 dill 0.3.5.1 dm-tree 0.1.7 einops 0.3.0 etils 0.6.0 flatbuffers 1.12 flax 0.5.2 flaxformer 0.4.2 fonttools 4.33.3 gast 0.4.0 google-auth 2.8.0 google-auth-oauthlib 0.4.6 google-pasta 0.2.0 googleapis-common-protos 1.56.3 grpcio 1.47.0 h5py 3.7.0 idna 3.3 importlib-metadata 4.12.0 importlib-resources 5.8.0 jax 0.3.14 jaxlib 0.3.14+cuda11.cudnn82 keras 2.9.0 Keras-Preprocessing 1.1.2 kiwisolver 1.4.3 libclang 14.0.1 Markdown 3.3.7 matplotlib 3.5.2 ml-collections 0.1.1 msgpack 1.0.4 numpy 1.21.6 oauthlib 3.2.0 opt-einsum 3.3.0 optax 0.1.2 packaging 21.3 pandas 1.3.5 Pillow 9.1.1 pip 21.2.2 promise 2.3 protobuf 3.19.4 pyasn1 0.4.8 pyasn1-modules 0.2.8 Pygments 2.12.0 pyparsing 3.0.9 python-dateutil 2.8.2 pytz 2022.1 PyYAML 6.0 requests 2.28.0 requests-oauthlib 1.3.1 rich 11.2.0 rsa 4.8 scipy 1.7.3 setuptools 61.2.0 six 1.16.0 tensorboard 2.9.1 tensorboard-data-server 0.6.1 tensorboard-plugin-wit 1.8.1 tensorflow-addons 0.17.1 tensorflow-cpu 2.9.1 tensorflow-datasets 4.6.0 tensorflow-estimator 2.9.0 tensorflow-hub 0.12.0 tensorflow-io-gcs-filesystem 0.26.0 tensorflow-metadata 1.9.0 tensorflow-probability 0.17.0 tensorflow-text 2.9.0 termcolor 1.1.0 toml 0.10.2 toolz 0.11.2 tqdm 4.64.0 typeguard 2.13.3 typing_extensions 4.2.0 urllib3 1.26.9 Werkzeug 2.1.2 wheel 0.37.1 wrapt 1.14.1 zipp 3.8.0

    opened by adamantboy 15
  • Question about exporting an integer-only MobileBERT to TF-Lite format.

    Question about exporting an integer-only MobileBERT to TF-Lite format.

    Hi, I'm trying to export a mobilebert model to tflite format.

    Environment Docker (tensorflow/tensorflow:1.15.0-gpu-py3) image V100 16GB

    As guided in README.md., I followed "Run Quantization-aware-training with Squad" then "Export an integer-only MobileBERT to TF-Lite format." However, I got an error while converting to quantized tflite model.

    2020-07-15 10:26:10.934857: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:786] Optimization results for grappler item: graph_to_optimize 2020-07-15 10:26:10.934903: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:788] constant_folding: Graph size after: 4461 nodes (-1120), 4701 edges (-1124), time = 779.203ms. 2020-07-15 10:26:10.934931: I tensorflow/core/grappler/optimizers/meta_optimizer.cc:788] constant_folding: Graph size after: 4461 nodes (0), 4701 edges (0), time = 374.792ms. Traceback (most recent call last): File "run_squad.py", line 1517, in tf.app.run() File "/usr/local/lib/python3.6/dist-packages/tensorflow_core/python/platform/app.py", line 40, in run _run(main=main, argv=argv, flags_parser=_parse_flags_tolerate_undef) File "/usr/local/lib/python3.6/dist-packages/absl/app.py", line 299, in run _run_main(main, args) File "/usr/local/lib/python3.6/dist-packages/absl/app.py", line 250, in _run_main sys.exit(main(argv)) File "run_squad.py", line 1508, in main tflite_model = converter.convert() File "/usr/local/lib/python3.6/dist-packages/tensorflow_core/lite/python/lite.py", line 993, in convert inference_output_type) File "/usr/local/lib/python3.6/dist-packages/tensorflow_core/lite/python/lite.py", line 239, in _calibrate_quantize_model inference_output_type, allow_float) File "/usr/local/lib/python3.6/dist-packages/tensorflow_core/lite/python/optimize/calibrator.py", line 78, in calibrate_and_quantize np.dtype(output_type.as_numpy_dtype()).num, allow_float) File "/usr/local/lib/python3.6/dist-packages/tensorflow_core/lite/python/optimize/tensorflow_lite_wrap_calibration_wrapper.py", line 115, in QuantizeModel return _tensorflow_lite_wrap_calibration_wrapper.CalibrationWrapper_QuantizeModel(self, input_py_type, output_py_type, allow_float) RuntimeError: Invalid quantization params for op GATHER at index 2 in subgraph 0

    I used pre-trained weights (uncased_L-24_H-128_B-512_A-4_F-4_OPT) that mentioned in README.md. Is it required to distillation process before quantization-aware-training?

    Regards, Dongjin.

    opened by nadongguri 13
  • kws_streaming.models.utils_test failure

    kws_streaming.models.utils_test failure

    I noticed that utils.model_to_tflite() is throwing errors with the GRU model. kws_streaming.models.utils_test fails as well with the following error. I observed this with tf2.1.0 as well as tf2.2.0-dev20200402.

    ERROR: test_model_to_tflite (__main__.UtilsTest)
    test_model_to_tflite (__main__.UtilsTest)
    TFLite supports stateless graphs.
    ----------------------------------------------------------------------
    Traceback (most recent call last):
      File "/Users/abhipray/NeoSensory/git-repos/google-research/kws_streaming/models/utils_test.py", line 85, in test_model_to_tflite
        self.assertTrue(utils.model_to_tflite(self.sess, self.model, self.flags))
      File "/Users/abhipray/NeoSensory/git-repos/google-research/kws_streaming/models/utils.py", line 295, in model_to_tflite
        sess, model_stateless_stream.inputs, model_stateless_stream.outputs)
      File "/anaconda3/lib/python3.7/site-packages/tensorflow/lite/python/lite.py", line 792, in from_session
        graph_def = _freeze_graph(sess, input_tensors, output_tensors)
      File "/anaconda3/lib/python3.7/site-packages/tensorflow/lite/python/util.py", line 256, in freeze_graph
        graph_def, input_tensors, output_tensors, config, graph=sess.graph)
      File "/anaconda3/lib/python3.7/site-packages/tensorflow/lite/python/util.py", line 218, in run_graph_optimizations
        return tf_optimizer.OptimizeGraph(config, meta_graph)
      File "/anaconda3/lib/python3.7/site-packages/tensorflow/python/grappler/tf_optimizer.py", line 58, in OptimizeGraph
        graph_id, strip_default_attributes)
    ValueError: Failed to import metagraph, check error log for more info.
    
    ----------------------------------------------------------------------
    Ran 5 tests in 2.938s```
    opened by Abhipray 13
  • Update Global Matplotlib Library

    Update Global Matplotlib Library

    Hello, i'm trying to use the current version of matplotlib 3.6.2 at colab, but some features are not avaliable at the colab instaled version wich is 3.2.2. So i've to upgrade matplot every time i'm using it pip install --upgrade matplotlib, why it's outdated?

    opened by AlexandroLuis 0
  • Bump setuptools from 58.2.0 to 65.5.1 in /nerflets

    Bump setuptools from 58.2.0 to 65.5.1 in /nerflets

    Bumps setuptools from 58.2.0 to 65.5.1.

    Release notes

    Sourced from setuptools's releases.

    v65.5.1

    No release notes provided.

    v65.5.0

    No release notes provided.

    v65.4.1

    No release notes provided.

    v65.4.0

    No release notes provided.

    v65.3.0

    No release notes provided.

    v65.2.0

    No release notes provided.

    v65.1.1

    No release notes provided.

    v65.1.0

    No release notes provided.

    v65.0.2

    No release notes provided.

    v65.0.1

    No release notes provided.

    v65.0.0

    No release notes provided.

    v64.0.3

    No release notes provided.

    v64.0.2

    No release notes provided.

    v64.0.1

    No release notes provided.

    v64.0.0

    No release notes provided.

    v63.4.3

    No release notes provided.

    v63.4.2

    No release notes provided.

    ... (truncated)

    Changelog

    Sourced from setuptools's changelog.

    v65.5.1

    Misc ^^^^

    • #3638: Drop a test dependency on the mock package, always use :external+python:py:mod:unittest.mock -- by :user:hroncok
    • #3659: Fixed REDoS vector in package_index.

    v65.5.0

    Changes ^^^^^^^

    • #3624: Fixed editable install for multi-module/no-package src-layout projects.
    • #3626: Minor refactorings to support distutils using stdlib logging module.

    Documentation changes ^^^^^^^^^^^^^^^^^^^^^

    • #3419: Updated the example version numbers to be compliant with PEP-440 on the "Specifying Your Project’s Version" page of the user guide.

    Misc ^^^^

    • #3569: Improved information about conflicting entries in the current working directory and editable install (in documentation and as an informational warning).
    • #3576: Updated version of validate_pyproject.

    v65.4.1

    Misc ^^^^

    • #3613: Fixed encoding errors in expand.StaticModule when system default encoding doesn't match expectations for source files.
    • #3617: Merge with pypa/distutils@6852b20 including fix for pypa/distutils#181.

    v65.4.0

    Changes ^^^^^^^

    v65.3.0

    ... (truncated)

    Commits

    Dependabot compatibility score

    Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


    Dependabot commands and options

    You can trigger Dependabot actions by commenting on this PR:

    • @dependabot rebase will rebase this PR
    • @dependabot recreate will recreate this PR, overwriting any edits that have been made to it
    • @dependabot merge will merge this PR after your CI passes on it
    • @dependabot squash and merge will squash and merge this PR after your CI passes on it
    • @dependabot cancel merge will cancel a previously requested merge and block automerging
    • @dependabot reopen will reopen this PR if it is closed
    • @dependabot close will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
    • @dependabot ignore this major version will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
    • @dependabot ignore this minor version will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
    • @dependabot ignore this dependency will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
    • @dependabot use these labels will set the current labels as the default for future PRs for this repo and language
    • @dependabot use these reviewers will set the current reviewers as the default for future PRs for this repo and language
    • @dependabot use these assignees will set the current assignees as the default for future PRs for this repo and language
    • @dependabot use this milestone will set the current milestone as the default for future PRs for this repo and language

    You can disable automated security fix PRs for this repo from the Security Alerts page.

    dependencies 
    opened by dependabot[bot] 0
  • > git fetch origin

    > git fetch origin

    git fetch origin

    git checkout -b fahadfa-patch-7 origin/fahadfa-patch-7

    git merge main

    #Originally posted by @fahadfa in https://jazeerapaints.com/issues/issuecomment-792187934

    __Originally posted by @fahadfa in https://github.com/fahadfa/jazeerapaint/[jazeerapaints.com/saudi-ar/issuecomment-792187934/]

    Originally posted by @fahadfa in https://github.com/Jazzeerapaints/-/issues/5#issuecomment-1129413662

    opened by fahadfa 0
  • Getting error 'Op type not registered 'Scann>ScannSearchBatched' while inferencing a TF recommender SCANN model in Java

    Getting error 'Op type not registered 'Scann>ScannSearchBatched' while inferencing a TF recommender SCANN model in Java

    Hi, I trained a ScaNN powered model described here in Python.

    Tf version: 2.7.0 Environment: Mac

    I am trying to infer it in JAVA using native tensorflow-core-platform version 0.4.0

    Java code:

            URL modelURL = Main.class.getClassLoader().getResource("model/1");
            String modelPath = Paths.get(modelURL.toURI()).toString();
    
            SavedModelBundle model = SavedModelBundle.load(modelPath, "serve");
    

    I am getting error Op type not registered 'Scann>ScannSearchBatched' in binary running on xx-MacBook-Pro-2.local. Make sure the Op and Kernel are registered in the binary running in this process. Note that if you are loading a saved graph which used ops from tf.contrib, accessing (e.g.) `tf.contrib.resampler` should be done before importing the graph, as contrib ops are lazily registered when the module is first accessed.

    Please let me know what's going on? I know Mac is not supported by ScaNN. Is there a way I can install it on mac for development puposes?. Also, I don't want to use Tf serving for inference. Can I use Tf native core platform in JAVA for Scann based model?

    opened by tripathysa 0
Owner
Google Research
Google Research
This repository contains the implementations related to the experiments of a set of publicly available datasets that are used in the time series forecasting research space.

TSForecasting This repository contains the implementations related to the experiments of a set of publicly available datasets that are used in the tim

Rakshitha Godahewa 80 Dec 30, 2022
[CVPR 2021] Released code for Counterfactual Zero-Shot and Open-Set Visual Recognition

Counterfactual Zero-Shot and Open-Set Visual Recognition This project provides implementations for our CVPR 2021 paper Counterfactual Zero-S

null 144 Dec 24, 2022
NP DRAW paper released code

NP-DRAW: A Non-Parametric Structured Latent Variable Model for Image Generation This repo contains the official implementation for the NP-DRAW paper.

ZENG Xiaohui 22 Mar 13, 2022
Code of PVTv2 is released! PVTv2 largely improves PVTv1 and works better than Swin Transformer with ImageNet-1K pre-training.

Updates (2020/06/21) Code of PVTv2 is released! PVTv2 largely improves PVTv1 and works better than Swin Transformer with ImageNet-1K pre-training. Pyr

null 1.3k Jan 4, 2023
Released code for Objects are Different: Flexible Monocular 3D Object Detection, CVPR21

MonoFlex Released code for Objects are Different: Flexible Monocular 3D Object Detection, CVPR21. Work in progress. Installation This repo is tested w

Yunpeng 169 Dec 6, 2022
[ICCV 2021] Released code for Causal Attention for Unbiased Visual Recognition

CaaM This repo contains the codes of training our CaaM on NICO/ImageNet9 dataset. Due to my recent limited bandwidth, this codebase is still messy, wh

Wang Tan 66 Dec 31, 2022
This repository contains an overview of important follow-up works based on the original Vision Transformer (ViT) by Google.

This repository contains an overview of important follow-up works based on the original Vision Transformer (ViT) by Google.

null 75 Dec 2, 2022
A Multi-modal Model Chinese Spell Checker Released on ACL2021.

ReaLiSe ReaLiSe is a multi-modal Chinese spell checking model. This the office code for the paper Read, Listen, and See: Leveraging Multimodal Informa

DaDa 106 Dec 29, 2022
Fuzzification helps developers protect the released, binary-only software from attackers who are capable of applying state-of-the-art fuzzing techniques

About Fuzzification Fuzzification helps developers protect the released, binary-only software from attackers who are capable of applying state-of-the-

gts3.org (SSLab@Gatech) 55 Oct 25, 2022
This is the first released system towards complex meters` detection and recognition, which is implemented by computer vision techniques.

A three-stage detection and recognition pipeline of complex meters in wild This is the first released system towards detection and recognition of comp

Yan Shu 19 Nov 28, 2022
FAIR's research platform for object detection research, implementing popular algorithms like Mask R-CNN and RetinaNet.

Detectron is deprecated. Please see detectron2, a ground-up rewrite of Detectron in PyTorch. Detectron Detectron is Facebook AI Research's software sy

Facebook Research 25.5k Jan 7, 2023
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)

MMF is a modular framework for vision and language multimodal research from Facebook AI Research. MMF contains reference implementations of state-of-t

Facebook Research 5.1k Jan 4, 2023
Pytorch port of Google Research's LEAF Audio paper

leaf-audio-pytorch Pytorch port of Google Research's LEAF Audio paper published at ICLR 2021. This port is not completely finished, but the Leaf() fro

Dennis Fedorishin 80 Oct 31, 2022
Research shows Google collects 20x more data from Android than Apple collects from iOS. Block this non-consensual telemetry using pihole blocklists.

pihole-antitelemetry Research shows Google collects 20x more data from Android than Apple collects from iOS. Block both using these pihole lists. Proj

Adrian Edwards 290 Jan 9, 2023
An image base contains 490 images for learning (400 cars and 90 boats), and another 21 images for testingAn image base contains 490 images for learning (400 cars and 90 boats), and another 21 images for testing

SVM Données Une base d’images contient 490 images pour l’apprentissage (400 voitures et 90 bateaux), et encore 21 images pour fait des tests. Prétrait

Achraf Rahouti 3 Nov 30, 2021
Red Team tool for exfiltrating files from a target's Google Drive that you have access to, via Google's API.

GD-Thief Red Team tool for exfiltrating files from a target's Google Drive that you(the attacker) has access to, via the Google Drive API. This includ

Antonio Piazza 39 Dec 27, 2022
A large dataset of 100k Google Satellite and matching Map images, resembling pix2pix's Google Maps dataset.

Larger Google Sat2Map dataset This dataset extends the aerial ⟷ Maps dataset used in pix2pix (Isola et al., CVPR17). The provide script download_sat2m

null 34 Dec 28, 2022
Google-drive-to-sqlite - Create a SQLite database containing metadata from Google Drive

google-drive-to-sqlite Create a SQLite database containing metadata from Google

Simon Willison 140 Dec 4, 2022
This repository contains the code used for Predicting Patient Outcomes with Graph Representation Learning (https://arxiv.org/abs/2101.03940).

Predicting Patient Outcomes with Graph Representation Learning This repository contains the code used for Predicting Patient Outcomes with Graph Repre

Emma Rocheteau 76 Dec 22, 2022