Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
WickedGoat98
Moderator
13 Questions, 96 Answers
  Active since 10 January 2023
  Last activity 12 days ago

Reputation

0

Badges 1

96 × Eureka!
0 Votes
0 Answers
552 Views
0 Votes 0 Answers 552 Views
Hi guys, my second article about ClearML is out 🙂 https://w-goutas.medium.com/track-your-experiments-with-clearml-in-docker-37c5bdb3c582 thanks for reading 😉
3 years ago
0 Votes
5 Answers
504 Views
0 Votes 5 Answers 504 Views
3 years ago
0 Votes
17 Answers
630 Views
0 Votes 17 Answers 630 Views
3 years ago
0 Votes
5 Answers
648 Views
0 Votes 5 Answers 648 Views
I cloned it and scheduled it to the default queue, but it is not being processed. Is the default queue by default not usable?
3 years ago
0 Votes
0 Answers
577 Views
0 Votes 0 Answers 577 Views
also the first run, even if it shows me its output and is done is marked still as running 😞
3 years ago
0 Votes
28 Answers
615 Views
0 Votes 28 Answers 615 Views
Hi guys, I managed to set up a Kubernetes cluster and install Trains into it. While testing my set-up I run the test_reporting.py example https://github.com/...
3 years ago
0 Votes
25 Answers
570 Views
0 Votes 25 Answers 570 Views
Hi, I expect there is a limitation in time the free service http://app.community.clear.ml is storing projects and experiments of an user. Can you please tell...
3 years ago
0 Votes
4 Answers
554 Views
0 Votes 4 Answers 554 Views
Hey guys, I just published (my second over all) article on Medium 😁 https://medium.com/@w.goutas/track-your-experiments-with-clearml-b26b2d0d6518
3 years ago
0 Votes
18 Answers
558 Views
0 Votes 18 Answers 558 Views
Hi all, I'm trying to deploy trains on Rancher (nice Kubernetes cluster orchestration project) where I'm quite new to Rancher and Kubernetes. I have been abl...
3 years ago
0 Votes
4 Answers
562 Views
0 Votes 4 Answers 562 Views
Hi, I'm using the dockerized version of trains get an understanding of trains. While trying to play with the trains.conf settings in ~/trains.conf I got in a...
3 years ago
0 Votes
23 Answers
588 Views
0 Votes 23 Answers 588 Views
3 years ago
0 Votes
26 Answers
535 Views
0 Votes 26 Answers 535 Views
3 years ago
0 Votes
5 Answers
646 Views
0 Votes 5 Answers 646 Views
3 years ago
0 Hi, I Have A Pre-Processing Steps Not Been Implemented In Python, But Being A Shell Script Calling Wget To Synchronize Data And Creating Intermediate Sqlite Dbs By A Script Been Implemented In 'R' And Would Like To Ask, If Trains Can Be Used Just To Trigg

need to read about the PipelineController. On a first view to the example it looks like what I would like to do.
I I would like to schedule multiple actions like 30 time the same script with different parameter, it looks like the add_step is what I will need

3 years ago
0 Hi, I Run The Trains Server In An Docker Container And Started Making Use Of Tasks ... My Tests Are Showed On The Projects Dashboard Which Is Realy Cool. What I Haven'T Found So Far Is A Way To Clean Up The System From The Tests I Did. I'M Able To Archive

Hi Martin,
you are right. The Trains-agent is running with option cpu-only
` (py38) wgo@NVidia-power:~/dev/catwalk$ docker ps
CONTAINER ID       IMAGE                                                COMMAND                 CREATED            STATUS             PORTS                                        
 NAMES
b99d5103a43c       allegroai/trains-agent-services:latest               "/usr/agent/entrypoi…"  2 days ago         Up 2 days                                                        
 ...

3 years ago
3 years ago
0 Hi, I Run The Trains Server In An Docker Container And Started Making Use Of Tasks ... My Tests Are Showed On The Projects Dashboard Which Is Realy Cool. What I Haven'T Found So Far Is A Way To Clean Up The System From The Tests I Did. I'M Able To Archive

I ran an local (not dockerized) trains-agent
trains-agent daemon --queue training --create-queue --foregroundwhich enabled me to see the GPU load on the corresponding view 🙂

Now I got another issue.
It seems when cloning an experiment, a virtual environment is been created with all the modules been identified to be used. Inside this environment the experiment is running.
Am I right?
Is this the case only for clones?

In my Python code I'm trying to read a pandas table which I stored i...

3 years ago
0 Hi, I Run The Trains Server In An Docker Container And Started Making Use Of Tasks ... My Tests Are Showed On The Projects Dashboard Which Is Realy Cool. What I Haven'T Found So Far Is A Way To Clean Up The System From The Tests I Did. I'M Able To Archive

after adding the
import fastparquet
statement to the code, the reconstruction of an clone is working
` Summary - installed python packages:
...

  • fastparquet==0.4.1
    ...
    Environment setup completed successfully
    Starting Task Execution:
    ...
    modeller.py: error: the following arguments are required: --algorithm `unfortunately it raises the next issue.
    If the script been used expects to get parameters via command line (which in Trains experiments are identified and stored as parameter when using...
3 years ago
3 years ago
0 Hi, I Run The Trains Server In An Docker Container And Started Making Use Of Tasks ... My Tests Are Showed On The Projects Dashboard Which Is Realy Cool. What I Haven'T Found So Far Is A Way To Clean Up The System From The Tests I Did. I'M Able To Archive

Sorry, but I don'T understand how the cloned experiment is been provided with parameters.
A task which is been cloned by Trains might get its parameter via task.set_parameters(dict)
this parameters are comming from soe magic analysis of the argparse been used in the script.
AgitatedDove14 when is the call to set_parameter(...) been performed? Is the argparse call been somehow redirected and will receive the data from Trains instead of getting them via sys.argv or wherever argparse is gettin...

3 years ago
0 Hi, I Run The Trains Server In An Docker Container And Started Making Use Of Tasks ... My Tests Are Showed On The Projects Dashboard Which Is Realy Cool. What I Haven'T Found So Far Is A Way To Clean Up The System From The Tests I Did. I'M Able To Archive

well I managed to clone an experiment and adat its parameter on the trains server via browser.
If argparse is been used, no parameter must be defined as required. Instead it has to be managed by the script after parsing the parameter and something mandatory is missing to terminate.
Doing so worked fine for me 😁 at least for this part of work. Now fastparquet and missing packages are failing again...

3 years ago
0 Hello Guys, I Read About Trains Some Days Ago And Think It Is Exectly What I Was Looking For, So I Ran The Docker Image And Started Thinking Of What I Would Like To Do And The Processing Steps I Would Like To Automize Which I Currently Run Manually Trigge

Hi Martin,

thanks for the reply.
The data I'm syncing by an data provider wich supports only an ftp connection....
I started a Ranche training, will need some time to be able to set-up my cluster before I can start using Trains ;)

3 years ago
0 Hi Guys, I Managed To Set Up A Kubernetes Cluster And Install Trains Into It. While Testing My Set-Up I Run The Test_Reporting.Py Example

to be honest, I don't know if I will find it as it is a Kubernetes cluster (ok only 2 nodes) and might be installed to somewhere ...
I will check if I will find any trains configs on the systems, but they should be the defaults comming with the Helm installer

3 years ago
0 Hi Guys, I Managed To Set Up A Kubernetes Cluster And Install Trains Into It. While Testing My Set-Up I Run The Test_Reporting.Py Example

` sdk {
# TRAINS - default SDK configuration

storage {
    cache {
        # Defaults to system temp folder / cache
        default_base_dir: "~/.trains/cache"
    }

    direct_access: [
        # Objects matching are considered to be available for direct access, i.e. they will not be downloaded
        # or cached, and any download request will return a direct reference.
        # Objects are specified in glob format, available for url and content_ty...
3 years ago
0 Hi Guys, I Managed To Set Up A Kubernetes Cluster And Install Trains Into It. While Testing My Set-Up I Run The Test_Reporting.Py Example

# TRAINS SDK configuration file api { # Notice: 'host' is the api server (default port 8008), not the web server. api_server: web_server: files_server: `
# Credentials are generated using the webapp, /profile
# Override with os environment: TRAINS_API_ACCESS_KEY / TRAINS_API_SECRET_KEY
credentials {....}
}
sdk {
# TRAINS - default SDK configuration

`

3 years ago
Show more results compactanswers