Resources Limit

Submissions

You can only submit during the Submission Phase.

You can only submit up to 5 times per day. Some competition may allow more. You can see your current submission for the day at the bottom left.

A submission cannot exceed 5GB and his resources/ directory cannot exceed 10GB.

circle-info

We can exceptionally allow you more as we understand that debugging why your code works locally but not in the cloud environment can be very difficult.

For this, please contact us on Discord.

Models

Also known as the resources/ directory, it allows you to carry state information over multiple runs, provided the total size is under 10 GB.

There is no limit on the number of files themselves, but if we detect abuse, we may introduce one that could break existing models.

All types of file are permitted, including but not limited to:

  • persisted models (.joblib),

  • model weights (.pkl),

  • configuration files,

  • ...

Compared to including them in your submission, those files can be modified and persisted across multiple runs.

Runs

For a run to be considered valid, it must:

  • not crash due to a bug in your code or because you have run out of RAM/disk space,

  • complete the work under the time constraints,

  • produce a prediction.

Predictions are checked once the run is over. However, if a prediction is deemed invalid, the run itself will not be invalidated.

Resources

You can view the runtime specifications before creating a run:

Runtime option selector on the Run creation page.
circle-info

The "Authorized Quota" is global; it is not allocated to each individual runtime.

Some consume the quota faster, some slower, but almost always at the normal speed. This rate is based on how much they cost and how powerful they are.

Providers

We have used AWS's managed Fargatearrow-up-right and Batcharrow-up-right (using g4dn.8xlarge instancesarrow-up-right) services for a long time, but we recently introduced a new provider: OpenGPUarrow-up-right, which provides faster and more powerful GPUs for when you require extra processing power.

Using one provider instead of another should not affect your code. If you think there is a bug, please contact us on Discord.

Networking

Access to the internet in a cloud environment is not permitted, but is handled differently depending on the type of runtime:

This difference in behaviour is due to the fact that GPUs rely on sockets for communication. The same applies to multiprocessing libraries: you need to choose a runtime with a GPU to make them work, even if you don't need the GPU itself.

Quota

The allocated quota are different per competition, you should the overview section to know how much is granted to every participants.

You can also visualize your quota usage under the Submissions & Runs tab.

Failed runs are not taken into account when computing the quota. This is to prevent situations where you accidentally use up your entire quota in one go and have to wait until next week to try again.

Resets

The quota resets every week during the Submission Phase.

It only resets once at the beginning of the Out-of-Sample Phase. Most competitions are unaffected by this, as they run the models on the entire dataset at once. However, some competitions run models on small sections of the dataset over several weeks.

Extras

Whenever you create a run, the timeout is set 30 minutes after your quota.

This allows you to still have a chance of completing the run if you only have a few minutes left.

This also covers the time necessary to set up the environment, such as downloading the data and installing the libraries. This is valuable time during which your model is not running, but which is necessary for the system.

But those minutes do not count towards your quota! If you exceed your quota -for example, if you use 15 hours and 30 seconds but your quota is 15 hours- you will not be able to create a new run.

Security

Your model runs with the lowest possible level of privilege.

Most of the time, the data is fed directly to your function as an argument, as some pre-processing operations (such as filtering) are sometimes needed.

Any users attempting to bypass the restriction mechanism will be disqualified!

Predictions

Your prediction must not exceed a certain size. This limit varies depending on the competition, but is usually large enough to accommodate everyone.

However, when participants are responsible for writing the prediction files themselves, they must also ensure that they name their files properly and use the correct flags to persist them in the correct format. The most common mistake is including the default pandas.DataFrame index when saving a CSV filearrow-up-right.

Participants cannot download prediction files. If a check fails, the error message should include enough details to help you debug the situation (extra or missing columns, NaN or infinite values, ...). If you need further assistance, please, contact us on Discord.

Last updated

Was this helpful?