Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Unanswered
Im Having Difficuilty Understanding How To Handle Modified Files On S3


ok, then, I have a solution, but it still makes duplicate names

  • new_dataset._dataset_link_entries = {} # Cleaning all raw/a.png files
  • resize a.png and save it in another location named a_resized.png
  • Add back other files i need (excluding raw/a.png), I add them to new_dataset._ dataset_link_entries
  • Use add_external_files to include it in dataset. Im also using dataset_path=[a list of relative paths]
    What I would expect:
    100 Files removed (all a.png)
    100 Files added (all a_resized.png)

What I get:

when doing new_dataset.list_files() it now returns me these double filenames: raw/a_resized.jpg/a_resized.jpg
Whats up with this?
Already checked all paths, i do not at any time pass double named files

  
  
Posted 4 months ago
63 Views
0 Answers
4 months ago
4 months ago