-
-
Notifications
You must be signed in to change notification settings - Fork 424
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Execution benchmarks now only take execution into account #431
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The benchmarks are failing because we define the same variables, maybe putting it in a block scope would help.
I'm personally ok with having the drop, there's enough of us with context as to what happened. When we merge to master we can mention why in the commit and it will show on the benchmarks page when you click through the commit hash |
Benchmark for 141db24Click to view benchmark
|
This is ready to go! :D |
The Symbol benchmark |
Benchmark for 141db24Click to view benchmark
|
Do you think we need some documentation here? |
I will add it. Do you think we should retroactively fix older benchmarks? This could be done by subtracting the Realm creation, the lexing and the parsing from the execution benchmarks programmatically. |
Codecov Report
@@ Coverage Diff @@
## master #431 +/- ##
=======================================
Coverage 66.00% 66.00%
=======================================
Files 146 146
Lines 9180 9180
=======================================
Hits 6059 6059
Misses 3121 3121 Continue to review full report at Codecov.
|
Benchmark for 7f0e4d9Click to view benchmark
|
Benchmark for f2496d8Click to view benchmark
|
You mean the benchmark graphs, so we don't get a misleading drop in benchmarks, right? If so, I think we should if it's possible. |
These benchmarks should be using the latest changes we merged yesterday in the master branch of criterion compare, right?
Yes, I meant that. I think it would be nicer. We could use the opportunity to normalize lexer + parser benchmarks. |
Yes. I was thinking about that too, I re-ran the benchmarks, the same thing happened. They should have been updated. 😕 Edit: maybe there is some kind of caching? |
Yeah if we can we should, do you have a plan of how to do this? |
Benchmark for f2496d8Click to view benchmark
|
Benchmark for f2496d8Click to view benchmark
|
Yep, I'm writing a script to calculate real execution benchmarks by subtracting realm execution + parsing + lexing, if available, if not, these two can be calculated approximately. |
I think this can be merged, right, @jasonwilliams? Probably before #458. |
Benchmark for f3a1ebeClick to view benchmark
|
This Pull Request tries to change the execution benchmarks in order to only benchmark the actual execution. This should give us much more interesting information on the actual execution.
To be determined:
I can also add benchmarks for #427, #429 and #430 if you'd like.