Better way to shuffle patches for image dataset- tf.data input pipeline












2















I have a image dataset of 4644 color images which i reshape to patches of size 50 x 50 and pass to my deep neural network.



The total no of patches which gets generated are 369765. I am using tf.data input pipeline for patches_generation.



My question is how to efficiently shuffle the patches before passing to network.



Is buffer size = 10000 in shuffle operation sufficient enough before passing to network or is there any other efficient way to shuffle among 369765 patches?



Steps that i followed:
1. Created a single tf-record which stores all 4644 images.
2. Use tf.data pipeline to decode each image and create patches from it.
3. Shuffle every 10000 patches and pass to network.



This is the code that i am using:
I am using buffer_size=10000, parallel_calls=4



dataset = (tf.data.TFRecordDataset( tfrecords_filename_image )
.repeat( no_epochs )
.map( read_and_decode, num_parallel_calls=num_parallel_calls )

.map( get_patches_fn, num_parallel_calls=num_parallel_calls )

.apply( tf.data.experimental.unbatch()) # unbatch the patches we just produced

.shuffle( buffer_size=buffer_size, seed=random_number_1 )
.batch( batch_size )
.prefetch( 1 )
)
get_patches_function definition:

get_patches_fn = lambda image: get_patches( image, patch_size=patch_size )

def get_patches( image, patch_size=16 ):
# Function to compute patches for given image
# Input- image - Image which has to be converted to patches
# patch_size- size of each patch
# Output-patches of image(4d Tensor)
# with tf.device('/cpu:0'):
pad = [ [ 0, 0 ], [ 0, 0 ] ]
patches_image = tf.space_to_batch_nd( [ image ], [ patch_size, patch_size ], pad )
patches_image = tf.split( patches_image, patch_size * patch_size, 0 )
patches_image = tf.stack( patches_image, 3 )
patches_image = tf.reshape( patches_image, [ -1, patch_size, patch_size, 3 ] )
)
return patches_image

read and decode function definition:

def read_and_decode( tf_record_file ):
# Function to read the tensorflow record and return image suitable for patching
# Input: tf_record_file - tf record file in which image can be extracted
# Output: Image

features = {
'height': tf.FixedLenFeature( [ ], tf.int64 ),
'width': tf.FixedLenFeature( [ ], tf.int64 ),
'image_raw': tf.FixedLenFeature( [ ], tf.string )
}
parsed = tf.parse_single_example( tf_record_file, features )
image = tf.decode_raw( parsed[ 'image_raw' ], tf.uint8 )
height = tf.cast( parsed[ 'height' ], tf.int32 )
width = tf.cast( parsed[ 'width' ], tf.int32 )
image_shape = tf.stack( [ height, width, -1 ] )
image = tf.reshape( image, image_shape )
image = image[ :, :, :3 ]
image = tf.cast( image, tf.float32 )

return image


Please also suggest if it's better to create separate tf-records for each images rather than a single tf-record for all images .



Thanks in Advance.










share|improve this question



























    2















    I have a image dataset of 4644 color images which i reshape to patches of size 50 x 50 and pass to my deep neural network.



    The total no of patches which gets generated are 369765. I am using tf.data input pipeline for patches_generation.



    My question is how to efficiently shuffle the patches before passing to network.



    Is buffer size = 10000 in shuffle operation sufficient enough before passing to network or is there any other efficient way to shuffle among 369765 patches?



    Steps that i followed:
    1. Created a single tf-record which stores all 4644 images.
    2. Use tf.data pipeline to decode each image and create patches from it.
    3. Shuffle every 10000 patches and pass to network.



    This is the code that i am using:
    I am using buffer_size=10000, parallel_calls=4



    dataset = (tf.data.TFRecordDataset( tfrecords_filename_image )
    .repeat( no_epochs )
    .map( read_and_decode, num_parallel_calls=num_parallel_calls )

    .map( get_patches_fn, num_parallel_calls=num_parallel_calls )

    .apply( tf.data.experimental.unbatch()) # unbatch the patches we just produced

    .shuffle( buffer_size=buffer_size, seed=random_number_1 )
    .batch( batch_size )
    .prefetch( 1 )
    )
    get_patches_function definition:

    get_patches_fn = lambda image: get_patches( image, patch_size=patch_size )

    def get_patches( image, patch_size=16 ):
    # Function to compute patches for given image
    # Input- image - Image which has to be converted to patches
    # patch_size- size of each patch
    # Output-patches of image(4d Tensor)
    # with tf.device('/cpu:0'):
    pad = [ [ 0, 0 ], [ 0, 0 ] ]
    patches_image = tf.space_to_batch_nd( [ image ], [ patch_size, patch_size ], pad )
    patches_image = tf.split( patches_image, patch_size * patch_size, 0 )
    patches_image = tf.stack( patches_image, 3 )
    patches_image = tf.reshape( patches_image, [ -1, patch_size, patch_size, 3 ] )
    )
    return patches_image

    read and decode function definition:

    def read_and_decode( tf_record_file ):
    # Function to read the tensorflow record and return image suitable for patching
    # Input: tf_record_file - tf record file in which image can be extracted
    # Output: Image

    features = {
    'height': tf.FixedLenFeature( [ ], tf.int64 ),
    'width': tf.FixedLenFeature( [ ], tf.int64 ),
    'image_raw': tf.FixedLenFeature( [ ], tf.string )
    }
    parsed = tf.parse_single_example( tf_record_file, features )
    image = tf.decode_raw( parsed[ 'image_raw' ], tf.uint8 )
    height = tf.cast( parsed[ 'height' ], tf.int32 )
    width = tf.cast( parsed[ 'width' ], tf.int32 )
    image_shape = tf.stack( [ height, width, -1 ] )
    image = tf.reshape( image, image_shape )
    image = image[ :, :, :3 ]
    image = tf.cast( image, tf.float32 )

    return image


    Please also suggest if it's better to create separate tf-records for each images rather than a single tf-record for all images .



    Thanks in Advance.










    share|improve this question

























      2












      2








      2








      I have a image dataset of 4644 color images which i reshape to patches of size 50 x 50 and pass to my deep neural network.



      The total no of patches which gets generated are 369765. I am using tf.data input pipeline for patches_generation.



      My question is how to efficiently shuffle the patches before passing to network.



      Is buffer size = 10000 in shuffle operation sufficient enough before passing to network or is there any other efficient way to shuffle among 369765 patches?



      Steps that i followed:
      1. Created a single tf-record which stores all 4644 images.
      2. Use tf.data pipeline to decode each image and create patches from it.
      3. Shuffle every 10000 patches and pass to network.



      This is the code that i am using:
      I am using buffer_size=10000, parallel_calls=4



      dataset = (tf.data.TFRecordDataset( tfrecords_filename_image )
      .repeat( no_epochs )
      .map( read_and_decode, num_parallel_calls=num_parallel_calls )

      .map( get_patches_fn, num_parallel_calls=num_parallel_calls )

      .apply( tf.data.experimental.unbatch()) # unbatch the patches we just produced

      .shuffle( buffer_size=buffer_size, seed=random_number_1 )
      .batch( batch_size )
      .prefetch( 1 )
      )
      get_patches_function definition:

      get_patches_fn = lambda image: get_patches( image, patch_size=patch_size )

      def get_patches( image, patch_size=16 ):
      # Function to compute patches for given image
      # Input- image - Image which has to be converted to patches
      # patch_size- size of each patch
      # Output-patches of image(4d Tensor)
      # with tf.device('/cpu:0'):
      pad = [ [ 0, 0 ], [ 0, 0 ] ]
      patches_image = tf.space_to_batch_nd( [ image ], [ patch_size, patch_size ], pad )
      patches_image = tf.split( patches_image, patch_size * patch_size, 0 )
      patches_image = tf.stack( patches_image, 3 )
      patches_image = tf.reshape( patches_image, [ -1, patch_size, patch_size, 3 ] )
      )
      return patches_image

      read and decode function definition:

      def read_and_decode( tf_record_file ):
      # Function to read the tensorflow record and return image suitable for patching
      # Input: tf_record_file - tf record file in which image can be extracted
      # Output: Image

      features = {
      'height': tf.FixedLenFeature( [ ], tf.int64 ),
      'width': tf.FixedLenFeature( [ ], tf.int64 ),
      'image_raw': tf.FixedLenFeature( [ ], tf.string )
      }
      parsed = tf.parse_single_example( tf_record_file, features )
      image = tf.decode_raw( parsed[ 'image_raw' ], tf.uint8 )
      height = tf.cast( parsed[ 'height' ], tf.int32 )
      width = tf.cast( parsed[ 'width' ], tf.int32 )
      image_shape = tf.stack( [ height, width, -1 ] )
      image = tf.reshape( image, image_shape )
      image = image[ :, :, :3 ]
      image = tf.cast( image, tf.float32 )

      return image


      Please also suggest if it's better to create separate tf-records for each images rather than a single tf-record for all images .



      Thanks in Advance.










      share|improve this question














      I have a image dataset of 4644 color images which i reshape to patches of size 50 x 50 and pass to my deep neural network.



      The total no of patches which gets generated are 369765. I am using tf.data input pipeline for patches_generation.



      My question is how to efficiently shuffle the patches before passing to network.



      Is buffer size = 10000 in shuffle operation sufficient enough before passing to network or is there any other efficient way to shuffle among 369765 patches?



      Steps that i followed:
      1. Created a single tf-record which stores all 4644 images.
      2. Use tf.data pipeline to decode each image and create patches from it.
      3. Shuffle every 10000 patches and pass to network.



      This is the code that i am using:
      I am using buffer_size=10000, parallel_calls=4



      dataset = (tf.data.TFRecordDataset( tfrecords_filename_image )
      .repeat( no_epochs )
      .map( read_and_decode, num_parallel_calls=num_parallel_calls )

      .map( get_patches_fn, num_parallel_calls=num_parallel_calls )

      .apply( tf.data.experimental.unbatch()) # unbatch the patches we just produced

      .shuffle( buffer_size=buffer_size, seed=random_number_1 )
      .batch( batch_size )
      .prefetch( 1 )
      )
      get_patches_function definition:

      get_patches_fn = lambda image: get_patches( image, patch_size=patch_size )

      def get_patches( image, patch_size=16 ):
      # Function to compute patches for given image
      # Input- image - Image which has to be converted to patches
      # patch_size- size of each patch
      # Output-patches of image(4d Tensor)
      # with tf.device('/cpu:0'):
      pad = [ [ 0, 0 ], [ 0, 0 ] ]
      patches_image = tf.space_to_batch_nd( [ image ], [ patch_size, patch_size ], pad )
      patches_image = tf.split( patches_image, patch_size * patch_size, 0 )
      patches_image = tf.stack( patches_image, 3 )
      patches_image = tf.reshape( patches_image, [ -1, patch_size, patch_size, 3 ] )
      )
      return patches_image

      read and decode function definition:

      def read_and_decode( tf_record_file ):
      # Function to read the tensorflow record and return image suitable for patching
      # Input: tf_record_file - tf record file in which image can be extracted
      # Output: Image

      features = {
      'height': tf.FixedLenFeature( [ ], tf.int64 ),
      'width': tf.FixedLenFeature( [ ], tf.int64 ),
      'image_raw': tf.FixedLenFeature( [ ], tf.string )
      }
      parsed = tf.parse_single_example( tf_record_file, features )
      image = tf.decode_raw( parsed[ 'image_raw' ], tf.uint8 )
      height = tf.cast( parsed[ 'height' ], tf.int32 )
      width = tf.cast( parsed[ 'width' ], tf.int32 )
      image_shape = tf.stack( [ height, width, -1 ] )
      image = tf.reshape( image, image_shape )
      image = image[ :, :, :3 ]
      image = tf.cast( image, tf.float32 )

      return image


      Please also suggest if it's better to create separate tf-records for each images rather than a single tf-record for all images .



      Thanks in Advance.







      python tensorflow deep-learning






      share|improve this question













      share|improve this question











      share|improve this question




      share|improve this question










      asked Nov 24 '18 at 8:59









      Venkatesh satagopanVenkatesh satagopan

      2815




      2815
























          1 Answer
          1






          active

          oldest

          votes


















          1














          A single tf-record file for all images is probably good enough given the number of images you have. If you have multiple disks you can try to split the file into one file per disk for higher throughput, but I don't think this should substantially slow a pipeline with the size of yours.



          Re the shuffle buffer size, that's an empirical question. A shuffle buffer as big as the dataset will give you true IID sampling; a smaller shuffle buffer will approximate it. Usually more randomness is better, but up to a point, so I recommend trying out a few different buffer sizes (assuming you can't have a buffer which fits the entire dataset) and see what works for you.






          share|improve this answer
























          • Thanks for the explanation. I tried with buffer_size of 50000 for shuffling and it gives me the expected result.

            – Venkatesh satagopan
            Dec 6 '18 at 19:47











          Your Answer






          StackExchange.ifUsing("editor", function () {
          StackExchange.using("externalEditor", function () {
          StackExchange.using("snippets", function () {
          StackExchange.snippets.init();
          });
          });
          }, "code-snippets");

          StackExchange.ready(function() {
          var channelOptions = {
          tags: "".split(" "),
          id: "1"
          };
          initTagRenderer("".split(" "), "".split(" "), channelOptions);

          StackExchange.using("externalEditor", function() {
          // Have to fire editor after snippets, if snippets enabled
          if (StackExchange.settings.snippets.snippetsEnabled) {
          StackExchange.using("snippets", function() {
          createEditor();
          });
          }
          else {
          createEditor();
          }
          });

          function createEditor() {
          StackExchange.prepareEditor({
          heartbeatType: 'answer',
          autoActivateHeartbeat: false,
          convertImagesToLinks: true,
          noModals: true,
          showLowRepImageUploadWarning: true,
          reputationToPostImages: 10,
          bindNavPrevention: true,
          postfix: "",
          imageUploader: {
          brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
          contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
          allowUrls: true
          },
          onDemand: true,
          discardSelector: ".discard-answer"
          ,immediatelyShowMarkdownHelp:true
          });


          }
          });














          draft saved

          draft discarded


















          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53456668%2fbetter-way-to-shuffle-patches-for-image-dataset-tf-data-input-pipeline%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown

























          1 Answer
          1






          active

          oldest

          votes








          1 Answer
          1






          active

          oldest

          votes









          active

          oldest

          votes






          active

          oldest

          votes









          1














          A single tf-record file for all images is probably good enough given the number of images you have. If you have multiple disks you can try to split the file into one file per disk for higher throughput, but I don't think this should substantially slow a pipeline with the size of yours.



          Re the shuffle buffer size, that's an empirical question. A shuffle buffer as big as the dataset will give you true IID sampling; a smaller shuffle buffer will approximate it. Usually more randomness is better, but up to a point, so I recommend trying out a few different buffer sizes (assuming you can't have a buffer which fits the entire dataset) and see what works for you.






          share|improve this answer
























          • Thanks for the explanation. I tried with buffer_size of 50000 for shuffling and it gives me the expected result.

            – Venkatesh satagopan
            Dec 6 '18 at 19:47
















          1














          A single tf-record file for all images is probably good enough given the number of images you have. If you have multiple disks you can try to split the file into one file per disk for higher throughput, but I don't think this should substantially slow a pipeline with the size of yours.



          Re the shuffle buffer size, that's an empirical question. A shuffle buffer as big as the dataset will give you true IID sampling; a smaller shuffle buffer will approximate it. Usually more randomness is better, but up to a point, so I recommend trying out a few different buffer sizes (assuming you can't have a buffer which fits the entire dataset) and see what works for you.






          share|improve this answer
























          • Thanks for the explanation. I tried with buffer_size of 50000 for shuffling and it gives me the expected result.

            – Venkatesh satagopan
            Dec 6 '18 at 19:47














          1












          1








          1







          A single tf-record file for all images is probably good enough given the number of images you have. If you have multiple disks you can try to split the file into one file per disk for higher throughput, but I don't think this should substantially slow a pipeline with the size of yours.



          Re the shuffle buffer size, that's an empirical question. A shuffle buffer as big as the dataset will give you true IID sampling; a smaller shuffle buffer will approximate it. Usually more randomness is better, but up to a point, so I recommend trying out a few different buffer sizes (assuming you can't have a buffer which fits the entire dataset) and see what works for you.






          share|improve this answer













          A single tf-record file for all images is probably good enough given the number of images you have. If you have multiple disks you can try to split the file into one file per disk for higher throughput, but I don't think this should substantially slow a pipeline with the size of yours.



          Re the shuffle buffer size, that's an empirical question. A shuffle buffer as big as the dataset will give you true IID sampling; a smaller shuffle buffer will approximate it. Usually more randomness is better, but up to a point, so I recommend trying out a few different buffer sizes (assuming you can't have a buffer which fits the entire dataset) and see what works for you.







          share|improve this answer












          share|improve this answer



          share|improve this answer










          answered Dec 3 '18 at 19:23









          Alexandre PassosAlexandre Passos

          4,2711917




          4,2711917













          • Thanks for the explanation. I tried with buffer_size of 50000 for shuffling and it gives me the expected result.

            – Venkatesh satagopan
            Dec 6 '18 at 19:47



















          • Thanks for the explanation. I tried with buffer_size of 50000 for shuffling and it gives me the expected result.

            – Venkatesh satagopan
            Dec 6 '18 at 19:47

















          Thanks for the explanation. I tried with buffer_size of 50000 for shuffling and it gives me the expected result.

          – Venkatesh satagopan
          Dec 6 '18 at 19:47





          Thanks for the explanation. I tried with buffer_size of 50000 for shuffling and it gives me the expected result.

          – Venkatesh satagopan
          Dec 6 '18 at 19:47




















          draft saved

          draft discarded




















































          Thanks for contributing an answer to Stack Overflow!


          • Please be sure to answer the question. Provide details and share your research!

          But avoid



          • Asking for help, clarification, or responding to other answers.

          • Making statements based on opinion; back them up with references or personal experience.


          To learn more, see our tips on writing great answers.




          draft saved


          draft discarded














          StackExchange.ready(
          function () {
          StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53456668%2fbetter-way-to-shuffle-patches-for-image-dataset-tf-data-input-pipeline%23new-answer', 'question_page');
          }
          );

          Post as a guest















          Required, but never shown





















































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown

































          Required, but never shown














          Required, but never shown












          Required, but never shown







          Required, but never shown







          Popular posts from this blog

          Costa Masnaga

          Fotorealismo

          Sidney Franklin