Skip to content

This issue was moved to a discussion.

You can continue the conversation there. Go to discussion →

Other ways to export metrics in hyperparameter tuning jobs #221

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
dennis-ec opened this issue Jun 8, 2018 · 2 comments
Closed

Other ways to export metrics in hyperparameter tuning jobs #221

dennis-ec opened this issue Jun 8, 2018 · 2 comments

Comments

@dennis-ec
Copy link

Hello,

First of all thank you for implementing Hyperparameter tuning jobs.
This seems like a really useful feature I'd like to use very often.

However, I don't understand why the metrics are collected via logs.
Using regex to search over the logs seems complicated.
The documentation is also confusing without showing an concrete output which will match the shown regex code.

Why are the metrics not gathered via an output json file (in case for custom code)?
E.g. metrics.json which could be created in /op/ml/ouput/

@dennis-ec dennis-ec changed the title Other ways to retrieve metrics in hyperparameter tuning jobs Other ways to export metrics in hyperparameter tuning jobs Jun 8, 2018
@dennis-ec
Copy link
Author

A little Update: This blog post shows a regex expression as example, which seems more reasonable:
'Validation-accuracy=([0-9\\.]+)'

Despite that, the approach to search over logs still seems very inconvenient. So the Feature request for an alternative method still stands.

@leopd
Copy link
Contributor

leopd commented Jun 13, 2018

Thanks for the feedback. To explain the "why", one of our goals was to enable hyperparameter tuning on training code that customers don't necessarily understand or control -- e.g. you download a learning algorithm from github or get something from a colleague, and you aren't necessarily comfortable or even able to change the code, but still want to tune it. Regex's solve this problem nicely. As such, it's compatible with nearly every piece of ML code ever written, although it does bring its own set of challenges like crafting the right regex.

I agree the documentation & examples can be better. I've developed some tools to help with testing & debugging the MetricDefinition regexes. I'll look into how we might release such tools publicly.

apacker pushed a commit to apacker/sagemaker-python-sdk that referenced this issue Nov 15, 2018
…ights_loss_functions-fixed

Add notebook for class weights and loss functions
@aws aws locked and limited conversation to collaborators May 19, 2021

This issue was moved to a discussion.

You can continue the conversation there. Go to discussion →

Projects
None yet
Development

No branches or pull requests

4 participants