Use all 24 Gb for one application on the K80 GPU

The Tesla K80 has 24 Gb but, as far as I understand, it is shared between two GK210 GPUs on the same card. So actually it’s a card with two 12 Gb GPUs.
But is it still possible to use all 24 Gb for one application fx training a large model in PyTorch or Keras?