Challenges with Chat Interfaces#
Some of our chats use live inference endpoints to provide you with real-time access to the underlying model. You can continue to use the /score
endpoint within the Challenge notebooks. Alternatively, a chat interface is available in the Crucible UI. This interface allows you to make direct calls to structure and observed inputs and outputs.
If you have experience with API testing tools like Postman, Insomnia, Bruno, Burp Suite, or Caido, we recommend using these tools to proxy your requests. This approach allows you to easily inspect, modify, and record inputs and outputs from challenge endpoints, enhancing your efficiency.
For this example, we'll use the Piece of Cake Challenge as a walkthrough. Hit the "Send" button to interact with your model.
We'll start by engaging with the model to determine if we can gain some clues on extracting the flag from this basic tutorial Challenge:
This Challenge is designed with beginners. It seems the model is telling us something here from some flaky guardrails. Let's try encoding the prompt:
It worked!
Well done.
Note that in some of these challenges, and based on the LLM's outputs designed for the Challenges, you may have to extract the flag from the system output and manually parse and submit the flag. Alternatively, the UI will reflect a "flag detected" message for you and automatically submit the flag. You can still upload this flag to the /submit-flag
endpoint through the notebooks, or alternatively add this into the Submit Flag field, for example:
Try Your First Challenge#
Now that you're familiar with Challenges, you're ready to try your first one. We've crafted four introductory Challenges to get you started: Bear 1, Bear 2, Bear 3, and Bear 4.
Learn more about how to Complete Your First Challenge.
Need any help or technical problems? Check out the troubleshooting guidance.